var/home/core/zuul-output/0000755000175000017500000000000015116040024014516 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015116052140015464 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000005046255315116052131017702 0ustar rootrootDec 09 15:31:44 crc systemd[1]: Starting Kubernetes Kubelet... Dec 09 15:31:44 crc restorecon[4754]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:44 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 15:31:45 crc restorecon[4754]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Dec 09 15:31:45 crc restorecon[4754]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Dec 09 15:31:45 crc kubenswrapper[4894]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 09 15:31:45 crc kubenswrapper[4894]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Dec 09 15:31:45 crc kubenswrapper[4894]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 09 15:31:45 crc kubenswrapper[4894]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 09 15:31:45 crc kubenswrapper[4894]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Dec 09 15:31:45 crc kubenswrapper[4894]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.949961 4894 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956726 4894 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956778 4894 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956789 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956796 4894 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956804 4894 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956813 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956819 4894 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956825 4894 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956830 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956836 4894 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956842 4894 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956848 4894 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956854 4894 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956859 4894 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956865 4894 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956870 4894 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956876 4894 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956881 4894 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956887 4894 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956892 4894 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956897 4894 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956903 4894 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956909 4894 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956926 4894 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956933 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956938 4894 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956946 4894 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956952 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956959 4894 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956964 4894 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956970 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956976 4894 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956981 4894 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956986 4894 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956992 4894 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.956997 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957004 4894 feature_gate.go:330] unrecognized feature gate: Example Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957010 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957017 4894 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957024 4894 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957032 4894 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957038 4894 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957043 4894 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957048 4894 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957053 4894 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957060 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957065 4894 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957071 4894 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957076 4894 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957081 4894 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957086 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957092 4894 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957097 4894 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957102 4894 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957107 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957112 4894 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957119 4894 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957125 4894 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957130 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957135 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957141 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957146 4894 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957152 4894 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957157 4894 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957162 4894 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957167 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957172 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957180 4894 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957185 4894 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957190 4894 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.957195 4894 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957785 4894 flags.go:64] FLAG: --address="0.0.0.0" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957805 4894 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957819 4894 flags.go:64] FLAG: --anonymous-auth="true" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957828 4894 flags.go:64] FLAG: --application-metrics-count-limit="100" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957837 4894 flags.go:64] FLAG: --authentication-token-webhook="false" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957844 4894 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957852 4894 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957861 4894 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957867 4894 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957873 4894 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957880 4894 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957887 4894 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957894 4894 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957901 4894 flags.go:64] FLAG: --cgroup-root="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957907 4894 flags.go:64] FLAG: --cgroups-per-qos="true" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957913 4894 flags.go:64] FLAG: --client-ca-file="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957919 4894 flags.go:64] FLAG: --cloud-config="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957930 4894 flags.go:64] FLAG: --cloud-provider="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957938 4894 flags.go:64] FLAG: --cluster-dns="[]" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957957 4894 flags.go:64] FLAG: --cluster-domain="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957965 4894 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957973 4894 flags.go:64] FLAG: --config-dir="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957979 4894 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957986 4894 flags.go:64] FLAG: --container-log-max-files="5" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.957994 4894 flags.go:64] FLAG: --container-log-max-size="10Mi" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958001 4894 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958007 4894 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958014 4894 flags.go:64] FLAG: --containerd-namespace="k8s.io" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958038 4894 flags.go:64] FLAG: --contention-profiling="false" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958045 4894 flags.go:64] FLAG: --cpu-cfs-quota="true" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958051 4894 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958058 4894 flags.go:64] FLAG: --cpu-manager-policy="none" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958064 4894 flags.go:64] FLAG: --cpu-manager-policy-options="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958073 4894 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958080 4894 flags.go:64] FLAG: --enable-controller-attach-detach="true" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958086 4894 flags.go:64] FLAG: --enable-debugging-handlers="true" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958092 4894 flags.go:64] FLAG: --enable-load-reader="false" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958099 4894 flags.go:64] FLAG: --enable-server="true" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958105 4894 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958118 4894 flags.go:64] FLAG: --event-burst="100" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958124 4894 flags.go:64] FLAG: --event-qps="50" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958130 4894 flags.go:64] FLAG: --event-storage-age-limit="default=0" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958137 4894 flags.go:64] FLAG: --event-storage-event-limit="default=0" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958143 4894 flags.go:64] FLAG: --eviction-hard="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958151 4894 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958157 4894 flags.go:64] FLAG: --eviction-minimum-reclaim="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958164 4894 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958180 4894 flags.go:64] FLAG: --eviction-soft="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958187 4894 flags.go:64] FLAG: --eviction-soft-grace-period="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958196 4894 flags.go:64] FLAG: --exit-on-lock-contention="false" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958203 4894 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958209 4894 flags.go:64] FLAG: --experimental-mounter-path="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958215 4894 flags.go:64] FLAG: --fail-cgroupv1="false" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958221 4894 flags.go:64] FLAG: --fail-swap-on="true" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958228 4894 flags.go:64] FLAG: --feature-gates="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958235 4894 flags.go:64] FLAG: --file-check-frequency="20s" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958241 4894 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958247 4894 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958254 4894 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958261 4894 flags.go:64] FLAG: --healthz-port="10248" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958267 4894 flags.go:64] FLAG: --help="false" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958273 4894 flags.go:64] FLAG: --hostname-override="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958279 4894 flags.go:64] FLAG: --housekeeping-interval="10s" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958285 4894 flags.go:64] FLAG: --http-check-frequency="20s" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958291 4894 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958297 4894 flags.go:64] FLAG: --image-credential-provider-config="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958303 4894 flags.go:64] FLAG: --image-gc-high-threshold="85" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958309 4894 flags.go:64] FLAG: --image-gc-low-threshold="80" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958315 4894 flags.go:64] FLAG: --image-service-endpoint="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958321 4894 flags.go:64] FLAG: --kernel-memcg-notification="false" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958327 4894 flags.go:64] FLAG: --kube-api-burst="100" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958333 4894 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958339 4894 flags.go:64] FLAG: --kube-api-qps="50" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958345 4894 flags.go:64] FLAG: --kube-reserved="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958351 4894 flags.go:64] FLAG: --kube-reserved-cgroup="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958357 4894 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958363 4894 flags.go:64] FLAG: --kubelet-cgroups="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958369 4894 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958375 4894 flags.go:64] FLAG: --lock-file="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958380 4894 flags.go:64] FLAG: --log-cadvisor-usage="false" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958388 4894 flags.go:64] FLAG: --log-flush-frequency="5s" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958396 4894 flags.go:64] FLAG: --log-json-info-buffer-size="0" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958406 4894 flags.go:64] FLAG: --log-json-split-stream="false" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958424 4894 flags.go:64] FLAG: --log-text-info-buffer-size="0" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958430 4894 flags.go:64] FLAG: --log-text-split-stream="false" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958437 4894 flags.go:64] FLAG: --logging-format="text" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958443 4894 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958449 4894 flags.go:64] FLAG: --make-iptables-util-chains="true" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958455 4894 flags.go:64] FLAG: --manifest-url="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958461 4894 flags.go:64] FLAG: --manifest-url-header="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958470 4894 flags.go:64] FLAG: --max-housekeeping-interval="15s" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958477 4894 flags.go:64] FLAG: --max-open-files="1000000" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958484 4894 flags.go:64] FLAG: --max-pods="110" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958490 4894 flags.go:64] FLAG: --maximum-dead-containers="-1" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958496 4894 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958502 4894 flags.go:64] FLAG: --memory-manager-policy="None" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958509 4894 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958515 4894 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958521 4894 flags.go:64] FLAG: --node-ip="192.168.126.11" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958527 4894 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958549 4894 flags.go:64] FLAG: --node-status-max-images="50" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958555 4894 flags.go:64] FLAG: --node-status-update-frequency="10s" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958561 4894 flags.go:64] FLAG: --oom-score-adj="-999" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958568 4894 flags.go:64] FLAG: --pod-cidr="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958573 4894 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958583 4894 flags.go:64] FLAG: --pod-manifest-path="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958589 4894 flags.go:64] FLAG: --pod-max-pids="-1" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958596 4894 flags.go:64] FLAG: --pods-per-core="0" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958602 4894 flags.go:64] FLAG: --port="10250" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958608 4894 flags.go:64] FLAG: --protect-kernel-defaults="false" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958614 4894 flags.go:64] FLAG: --provider-id="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958621 4894 flags.go:64] FLAG: --qos-reserved="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958627 4894 flags.go:64] FLAG: --read-only-port="10255" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958633 4894 flags.go:64] FLAG: --register-node="true" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958661 4894 flags.go:64] FLAG: --register-schedulable="true" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958668 4894 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958687 4894 flags.go:64] FLAG: --registry-burst="10" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958692 4894 flags.go:64] FLAG: --registry-qps="5" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958698 4894 flags.go:64] FLAG: --reserved-cpus="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958712 4894 flags.go:64] FLAG: --reserved-memory="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958721 4894 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958728 4894 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958735 4894 flags.go:64] FLAG: --rotate-certificates="false" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958741 4894 flags.go:64] FLAG: --rotate-server-certificates="false" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958747 4894 flags.go:64] FLAG: --runonce="false" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958753 4894 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958760 4894 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958777 4894 flags.go:64] FLAG: --seccomp-default="false" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958783 4894 flags.go:64] FLAG: --serialize-image-pulls="true" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958789 4894 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958796 4894 flags.go:64] FLAG: --storage-driver-db="cadvisor" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958803 4894 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958810 4894 flags.go:64] FLAG: --storage-driver-password="root" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958816 4894 flags.go:64] FLAG: --storage-driver-secure="false" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958822 4894 flags.go:64] FLAG: --storage-driver-table="stats" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958828 4894 flags.go:64] FLAG: --storage-driver-user="root" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958835 4894 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958842 4894 flags.go:64] FLAG: --sync-frequency="1m0s" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958850 4894 flags.go:64] FLAG: --system-cgroups="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958856 4894 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958869 4894 flags.go:64] FLAG: --system-reserved-cgroup="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958875 4894 flags.go:64] FLAG: --tls-cert-file="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958881 4894 flags.go:64] FLAG: --tls-cipher-suites="[]" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958895 4894 flags.go:64] FLAG: --tls-min-version="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958901 4894 flags.go:64] FLAG: --tls-private-key-file="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958907 4894 flags.go:64] FLAG: --topology-manager-policy="none" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958914 4894 flags.go:64] FLAG: --topology-manager-policy-options="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958921 4894 flags.go:64] FLAG: --topology-manager-scope="container" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958927 4894 flags.go:64] FLAG: --v="2" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958941 4894 flags.go:64] FLAG: --version="false" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958949 4894 flags.go:64] FLAG: --vmodule="" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958957 4894 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.958963 4894 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959135 4894 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959142 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959156 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959162 4894 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959168 4894 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959174 4894 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959182 4894 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959187 4894 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959192 4894 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959197 4894 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959203 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959208 4894 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959213 4894 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959218 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959225 4894 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959232 4894 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959238 4894 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959243 4894 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959248 4894 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959254 4894 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959259 4894 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959264 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959270 4894 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959275 4894 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959280 4894 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959289 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959295 4894 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959300 4894 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959305 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959312 4894 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959319 4894 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959325 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959332 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959338 4894 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959345 4894 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959351 4894 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959358 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959364 4894 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959383 4894 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959391 4894 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959397 4894 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959404 4894 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959409 4894 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959416 4894 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959421 4894 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959428 4894 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959434 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959440 4894 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959447 4894 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959482 4894 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959490 4894 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959496 4894 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959502 4894 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959508 4894 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959513 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959518 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959523 4894 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959531 4894 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959536 4894 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959541 4894 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959547 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959552 4894 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959558 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959564 4894 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959571 4894 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959577 4894 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959582 4894 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959588 4894 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959594 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959599 4894 feature_gate.go:330] unrecognized feature gate: Example Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.959606 4894 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.959616 4894 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.968486 4894 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.968535 4894 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968620 4894 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968629 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968657 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968662 4894 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968666 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968669 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968674 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968677 4894 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968682 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968686 4894 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968690 4894 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968695 4894 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968699 4894 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968705 4894 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968710 4894 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968713 4894 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968717 4894 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968720 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968724 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968728 4894 feature_gate.go:330] unrecognized feature gate: Example Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968731 4894 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968735 4894 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968739 4894 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968743 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968747 4894 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968751 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968754 4894 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968758 4894 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968761 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968765 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968768 4894 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968772 4894 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968776 4894 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968780 4894 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968788 4894 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968792 4894 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968796 4894 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968799 4894 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968803 4894 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968807 4894 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968810 4894 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968813 4894 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968817 4894 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968821 4894 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968824 4894 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968829 4894 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968832 4894 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968836 4894 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968839 4894 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968843 4894 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968846 4894 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968851 4894 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968856 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968861 4894 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968865 4894 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968870 4894 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968874 4894 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968878 4894 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968882 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968887 4894 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968892 4894 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968895 4894 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968899 4894 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968903 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968906 4894 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968910 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968913 4894 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968917 4894 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968920 4894 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968925 4894 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.968930 4894 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.968937 4894 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969054 4894 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969060 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969064 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969068 4894 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969073 4894 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969078 4894 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969084 4894 feature_gate.go:330] unrecognized feature gate: NewOLM Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969089 4894 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969093 4894 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969097 4894 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969102 4894 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969106 4894 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969110 4894 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969115 4894 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969119 4894 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969123 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969127 4894 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969131 4894 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969135 4894 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969139 4894 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969143 4894 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969146 4894 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969153 4894 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969158 4894 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969163 4894 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969166 4894 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969170 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969174 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969178 4894 feature_gate.go:330] unrecognized feature gate: PlatformOperators Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969181 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969185 4894 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969189 4894 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969193 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfig Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969197 4894 feature_gate.go:330] unrecognized feature gate: GatewayAPI Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969201 4894 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969205 4894 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969209 4894 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969212 4894 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969218 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969221 4894 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969225 4894 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969228 4894 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969233 4894 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969237 4894 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969241 4894 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969245 4894 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969249 4894 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969252 4894 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969256 4894 feature_gate.go:330] unrecognized feature gate: PinnedImages Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969260 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969263 4894 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969266 4894 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969270 4894 feature_gate.go:330] unrecognized feature gate: SignatureStores Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969273 4894 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969277 4894 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969281 4894 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969284 4894 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969287 4894 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969291 4894 feature_gate.go:330] unrecognized feature gate: OVNObservability Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969296 4894 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969300 4894 feature_gate.go:330] unrecognized feature gate: Example Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969304 4894 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969308 4894 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969312 4894 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969316 4894 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969319 4894 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969323 4894 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969326 4894 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969331 4894 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969335 4894 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Dec 09 15:31:45 crc kubenswrapper[4894]: W1209 15:31:45.969340 4894 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.969346 4894 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.969753 4894 server.go:940] "Client rotation is on, will bootstrap in background" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.972211 4894 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.972290 4894 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.972777 4894 server.go:997] "Starting client certificate rotation" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.972803 4894 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.973136 4894 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-24 23:10:11.595901468 +0000 UTC Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.973318 4894 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 367h38m25.622587549s for next certificate rotation Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.982152 4894 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.983835 4894 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 09 15:31:45 crc kubenswrapper[4894]: I1209 15:31:45.995925 4894 log.go:25] "Validated CRI v1 runtime API" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.016273 4894 log.go:25] "Validated CRI v1 image API" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.018096 4894 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.023925 4894 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-12-09-15-27-24-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.023964 4894 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.043303 4894 manager.go:217] Machine: {Timestamp:2025-12-09 15:31:46.040732019 +0000 UTC m=+0.359942708 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654124544 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:90f20501-2d99-4429-ad71-02ebb56cf15c BootID:a485cf49-82ef-4626-a019-8dbcfdbd7ac7 Filesystems:[{Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365408768 Type:vfs Inodes:821633 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108169 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827060224 Type:vfs Inodes:4108169 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:4f:4c:4e Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:4f:4c:4e Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:77:5a:27 Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:5f:a9:e6 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:9c:a6:6a Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:ba:c8:68 Speed:-1 Mtu:1496} {Name:ens7.23 MacAddress:52:54:00:6f:4b:38 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:32:98:4e:c5:8a:95 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:f6:d0:cb:79:9c:22 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654124544 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.043575 4894 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.043816 4894 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.044446 4894 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.044687 4894 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.044731 4894 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.044953 4894 topology_manager.go:138] "Creating topology manager with none policy" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.044965 4894 container_manager_linux.go:303] "Creating device plugin manager" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.045160 4894 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.045194 4894 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.045344 4894 state_mem.go:36] "Initialized new in-memory state store" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.045726 4894 server.go:1245] "Using root directory" path="/var/lib/kubelet" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.046532 4894 kubelet.go:418] "Attempting to sync node with API server" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.046555 4894 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.046582 4894 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.046599 4894 kubelet.go:324] "Adding apiserver pod source" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.046613 4894 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.048659 4894 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.049181 4894 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.050863 4894 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Dec 09 15:31:46 crc kubenswrapper[4894]: W1209 15:31:46.051179 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.64:6443: connect: connection refused Dec 09 15:31:46 crc kubenswrapper[4894]: W1209 15:31:46.051290 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.64:6443: connect: connection refused Dec 09 15:31:46 crc kubenswrapper[4894]: E1209 15:31:46.051355 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.64:6443: connect: connection refused" logger="UnhandledError" Dec 09 15:31:46 crc kubenswrapper[4894]: E1209 15:31:46.051404 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.64:6443: connect: connection refused" logger="UnhandledError" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.051539 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.051570 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.051580 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.051590 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.051607 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.051617 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.051626 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.051654 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.051666 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.051677 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.051710 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.051720 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.052507 4894 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.053358 4894 server.go:1280] "Started kubelet" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.054304 4894 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.054303 4894 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.055345 4894 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.64:6443: connect: connection refused Dec 09 15:31:46 crc systemd[1]: Started Kubernetes Kubelet. Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.056207 4894 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.056425 4894 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.056457 4894 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.056510 4894 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-03 16:12:26.476022356 +0000 UTC Dec 09 15:31:46 crc kubenswrapper[4894]: E1209 15:31:46.058070 4894 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.059539 4894 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.059774 4894 factory.go:55] Registering systemd factory Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.059909 4894 factory.go:221] Registration of the systemd container factory successfully Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.060279 4894 volume_manager.go:287] "The desired_state_of_world populator starts" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.060360 4894 volume_manager.go:289] "Starting Kubelet Volume Manager" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.060588 4894 factory.go:153] Registering CRI-O factory Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.060607 4894 factory.go:221] Registration of the crio container factory successfully Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.060720 4894 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.060759 4894 factory.go:103] Registering Raw factory Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.060786 4894 manager.go:1196] Started watching for new ooms in manager Dec 09 15:31:46 crc kubenswrapper[4894]: E1209 15:31:46.061246 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" interval="200ms" Dec 09 15:31:46 crc kubenswrapper[4894]: W1209 15:31:46.064406 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.64:6443: connect: connection refused Dec 09 15:31:46 crc kubenswrapper[4894]: E1209 15:31:46.064491 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.64:6443: connect: connection refused" logger="UnhandledError" Dec 09 15:31:46 crc kubenswrapper[4894]: E1209 15:31:46.061276 4894 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.64:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187f95d18452ac88 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-09 15:31:46.053237896 +0000 UTC m=+0.372448565,LastTimestamp:2025-12-09 15:31:46.053237896 +0000 UTC m=+0.372448565,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.064558 4894 server.go:460] "Adding debug handlers to kubelet server" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.065095 4894 manager.go:319] Starting recovery of all containers Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.072173 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.072540 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.072775 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.072855 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.072979 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.073086 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.073171 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.073260 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.073368 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.073470 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.073564 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.073670 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.073754 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.073844 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.073943 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.074044 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.074135 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.074221 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.074311 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.074402 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.075380 4894 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.075507 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.075623 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.075744 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.075842 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.075939 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.076024 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.076111 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.076193 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.076279 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.076368 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.076459 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.076560 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.076662 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.076774 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.076876 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.076983 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.077078 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.077177 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.077267 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.077394 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.077487 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.077566 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.077675 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.077775 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.077863 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.077957 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.078040 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.078177 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.078271 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.078359 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.078503 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.078584 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.078663 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.078746 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.078837 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.079030 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.079125 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.079240 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.079324 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.079412 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.079510 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.079602 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.079713 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.079804 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.079893 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.079981 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.080082 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.080160 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.080255 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.080325 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.080389 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.080450 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.080516 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.080579 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.080661 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.080734 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.080806 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.080880 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.080943 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.081008 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.081077 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.081136 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.081201 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.081811 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.081909 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.082009 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.082094 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.082288 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.082389 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.082482 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.082596 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.082701 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.082781 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.082858 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.082925 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.082999 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.083087 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.083205 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.083302 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.083376 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.083439 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.083504 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.083569 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.083721 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.086543 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.086685 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.086722 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.086756 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.086789 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.086815 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.086846 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.086879 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.086902 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.086938 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.086961 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.086991 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087016 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087037 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087064 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087083 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087110 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087135 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087191 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087220 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087297 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087358 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087423 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087489 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087516 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087548 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087569 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087589 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087616 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087672 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087701 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087723 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087740 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087760 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087777 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087798 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087840 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087858 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087878 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087895 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087916 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087932 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.087961 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.088001 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.088020 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.088050 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.088072 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.088090 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.088110 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.088130 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.088153 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.088807 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.088842 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.088858 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.088873 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.088888 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.088903 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.088918 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.088931 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.088945 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.088962 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.088977 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.088991 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089007 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089023 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089037 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089051 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089064 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089080 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089096 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089110 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089126 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089140 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089154 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089169 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089184 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089198 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089212 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089228 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089263 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089278 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089291 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089307 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089336 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089353 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089367 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089396 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089412 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089448 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089477 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089495 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089512 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089537 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089552 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089576 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089596 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089613 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089629 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089667 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089687 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089705 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089722 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089739 4894 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089755 4894 reconstruct.go:97] "Volume reconstruction finished" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.089765 4894 reconciler.go:26] "Reconciler: start to sync state" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.092824 4894 manager.go:324] Recovery completed Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.102892 4894 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.105055 4894 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.105124 4894 status_manager.go:217] "Starting to sync pod status with apiserver" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.105163 4894 kubelet.go:2335] "Starting kubelet main sync loop" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.105184 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:46 crc kubenswrapper[4894]: E1209 15:31:46.105215 4894 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Dec 09 15:31:46 crc kubenswrapper[4894]: W1209 15:31:46.106024 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.64:6443: connect: connection refused Dec 09 15:31:46 crc kubenswrapper[4894]: E1209 15:31:46.106896 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.64:6443: connect: connection refused" logger="UnhandledError" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.107478 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.107515 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.107523 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.108903 4894 cpu_manager.go:225] "Starting CPU manager" policy="none" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.108923 4894 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.108940 4894 state_mem.go:36] "Initialized new in-memory state store" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.118354 4894 policy_none.go:49] "None policy: Start" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.121667 4894 memory_manager.go:170] "Starting memorymanager" policy="None" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.121697 4894 state_mem.go:35] "Initializing new in-memory state store" Dec 09 15:31:46 crc kubenswrapper[4894]: E1209 15:31:46.155466 4894 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.64:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187f95d18452ac88 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-09 15:31:46.053237896 +0000 UTC m=+0.372448565,LastTimestamp:2025-12-09 15:31:46.053237896 +0000 UTC m=+0.372448565,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 09 15:31:46 crc kubenswrapper[4894]: E1209 15:31:46.158744 4894 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.167506 4894 manager.go:334] "Starting Device Plugin manager" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.167560 4894 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.167574 4894 server.go:79] "Starting device plugin registration server" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.168004 4894 eviction_manager.go:189] "Eviction manager: starting control loop" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.168023 4894 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.168181 4894 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.168283 4894 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.168310 4894 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Dec 09 15:31:46 crc kubenswrapper[4894]: E1209 15:31:46.175307 4894 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.205463 4894 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.205608 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.206823 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.206907 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.206940 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.207129 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.207357 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.207406 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.208060 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.208083 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.208091 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.208253 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.208284 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.208295 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.208305 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.208337 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.208257 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.208989 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.209012 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.209037 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.209254 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.209276 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.209286 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.209383 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.209476 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.209525 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.209931 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.209947 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.209954 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.210022 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.210114 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.210155 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.210123 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.210231 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.210243 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.210488 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.210517 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.210527 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.210721 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.210757 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.210756 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.210920 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.210935 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.211686 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.211719 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.211735 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:46 crc kubenswrapper[4894]: E1209 15:31:46.262258 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" interval="400ms" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.268465 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.269516 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.269554 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.269565 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.269593 4894 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 15:31:46 crc kubenswrapper[4894]: E1209 15:31:46.269985 4894 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.64:6443: connect: connection refused" node="crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.291556 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.291617 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.291672 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.291707 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.291768 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.291944 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.292075 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.292172 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.292204 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.292241 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.292265 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.292341 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.292422 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.292465 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.292512 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.393250 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.393314 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.393346 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.393376 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.393406 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.393434 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.393461 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.393485 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.393506 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.393535 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.393560 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.393584 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.393609 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.394275 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.394286 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.394294 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.394238 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.394470 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.393845 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.394086 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.393961 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.394314 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.393618 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.394004 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.394411 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.393845 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.394683 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.394772 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.394719 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.394925 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.470946 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.473340 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.473413 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.473428 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.473470 4894 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 15:31:46 crc kubenswrapper[4894]: E1209 15:31:46.474521 4894 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.64:6443: connect: connection refused" node="crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.550038 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.557483 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.583383 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: W1209 15:31:46.596871 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-6be79be4b7b8c64be5e1532a8d97f7e169681b3b897da2aa0ef9dc6f9bfdd324 WatchSource:0}: Error finding container 6be79be4b7b8c64be5e1532a8d97f7e169681b3b897da2aa0ef9dc6f9bfdd324: Status 404 returned error can't find the container with id 6be79be4b7b8c64be5e1532a8d97f7e169681b3b897da2aa0ef9dc6f9bfdd324 Dec 09 15:31:46 crc kubenswrapper[4894]: W1209 15:31:46.598618 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-94f147169beb507f764201c80c71180a3a21a7358aa16f6ba6890ac094c164d7 WatchSource:0}: Error finding container 94f147169beb507f764201c80c71180a3a21a7358aa16f6ba6890ac094c164d7: Status 404 returned error can't find the container with id 94f147169beb507f764201c80c71180a3a21a7358aa16f6ba6890ac094c164d7 Dec 09 15:31:46 crc kubenswrapper[4894]: W1209 15:31:46.603388 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-dfa00f10762ff8f48497da5edc401e9d1e222ca107beea337f815d2cc5e599d0 WatchSource:0}: Error finding container dfa00f10762ff8f48497da5edc401e9d1e222ca107beea337f815d2cc5e599d0: Status 404 returned error can't find the container with id dfa00f10762ff8f48497da5edc401e9d1e222ca107beea337f815d2cc5e599d0 Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.612130 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.617256 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 15:31:46 crc kubenswrapper[4894]: W1209 15:31:46.633909 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-ef92b595a4a56a8122744ad44eac5dcd27d544c72e285043ff2fdb318493d4f6 WatchSource:0}: Error finding container ef92b595a4a56a8122744ad44eac5dcd27d544c72e285043ff2fdb318493d4f6: Status 404 returned error can't find the container with id ef92b595a4a56a8122744ad44eac5dcd27d544c72e285043ff2fdb318493d4f6 Dec 09 15:31:46 crc kubenswrapper[4894]: E1209 15:31:46.663610 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" interval="800ms" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.875080 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.876234 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.876279 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.876290 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:46 crc kubenswrapper[4894]: I1209 15:31:46.876314 4894 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 15:31:46 crc kubenswrapper[4894]: E1209 15:31:46.876744 4894 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.64:6443: connect: connection refused" node="crc" Dec 09 15:31:47 crc kubenswrapper[4894]: W1209 15:31:47.023128 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.64:6443: connect: connection refused Dec 09 15:31:47 crc kubenswrapper[4894]: E1209 15:31:47.023603 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.64:6443: connect: connection refused" logger="UnhandledError" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.057023 4894 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-23 12:43:09.748802608 +0000 UTC Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.057110 4894 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 333h11m22.691694752s for next certificate rotation Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.057211 4894 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.64:6443: connect: connection refused Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.112557 4894 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="8dc2c93cdf9eebdc700c2752e395e8c2d02279ca90ebbd528b1b594901559231" exitCode=0 Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.112692 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"8dc2c93cdf9eebdc700c2752e395e8c2d02279ca90ebbd528b1b594901559231"} Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.112787 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"6be79be4b7b8c64be5e1532a8d97f7e169681b3b897da2aa0ef9dc6f9bfdd324"} Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.112928 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.114102 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.114138 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.114148 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.114460 4894 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="210a487b2bd321692484f3a44391086b4062bff9e34d2391f10d9f83a6ee4696" exitCode=0 Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.114523 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"210a487b2bd321692484f3a44391086b4062bff9e34d2391f10d9f83a6ee4696"} Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.114550 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"94f147169beb507f764201c80c71180a3a21a7358aa16f6ba6890ac094c164d7"} Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.114724 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.115499 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.115531 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.115544 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.116317 4894 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="771f15a0047608775ea7c49c6fd52f868423765c9ba57a43647c830947bb0ee8" exitCode=0 Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.116358 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"771f15a0047608775ea7c49c6fd52f868423765c9ba57a43647c830947bb0ee8"} Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.116373 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"3b12253a9f27047c724a2b43e353ad5407ec2b7a472e0baa3d15cb72d875481d"} Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.116436 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.117234 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.117247 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.117255 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.119558 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755"} Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.119616 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ef92b595a4a56a8122744ad44eac5dcd27d544c72e285043ff2fdb318493d4f6"} Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.121200 4894 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6" exitCode=0 Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.121219 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6"} Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.121271 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"dfa00f10762ff8f48497da5edc401e9d1e222ca107beea337f815d2cc5e599d0"} Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.121399 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.122500 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.122531 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.122541 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.125527 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.126608 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.126676 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.126687 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:47 crc kubenswrapper[4894]: W1209 15:31:47.272130 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.64:6443: connect: connection refused Dec 09 15:31:47 crc kubenswrapper[4894]: E1209 15:31:47.272202 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.64:6443: connect: connection refused" logger="UnhandledError" Dec 09 15:31:47 crc kubenswrapper[4894]: W1209 15:31:47.329160 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.64:6443: connect: connection refused Dec 09 15:31:47 crc kubenswrapper[4894]: E1209 15:31:47.329240 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.64:6443: connect: connection refused" logger="UnhandledError" Dec 09 15:31:47 crc kubenswrapper[4894]: W1209 15:31:47.334995 4894 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.64:6443: connect: connection refused Dec 09 15:31:47 crc kubenswrapper[4894]: E1209 15:31:47.335155 4894 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.64:6443: connect: connection refused" logger="UnhandledError" Dec 09 15:31:47 crc kubenswrapper[4894]: E1209 15:31:47.465102 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" interval="1.6s" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.677521 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.679502 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.679552 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.679564 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:47 crc kubenswrapper[4894]: I1209 15:31:47.679590 4894 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 15:31:47 crc kubenswrapper[4894]: E1209 15:31:47.680140 4894 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.64:6443: connect: connection refused" node="crc" Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.056775 4894 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.64:6443: connect: connection refused Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.127004 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2"} Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.127051 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295"} Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.127062 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10"} Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.127075 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca"} Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.128628 4894 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="58c502f898c98baaf18a50d794cbccb7d59ec1d41fcbd968e25432656712cac9" exitCode=0 Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.128714 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"58c502f898c98baaf18a50d794cbccb7d59ec1d41fcbd968e25432656712cac9"} Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.128846 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.131993 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.132022 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.132032 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.135299 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"fd9840b18660dd5f9a52708ce51bc209f5dd295b3b830019c0f9b20f6ec63618"} Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.135350 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.136227 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.136258 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.136269 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.143890 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b70791118fefd6a669a2e12536d1942d7120c7759a53e131f889242881918821"} Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.143964 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"882ae42ef0bd79cba2c436b8eeda122b33e1ed9efe5d09af3c80b3b1145c8103"} Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.144011 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"5293b95acee426113021da42be9937be7bcefa80361ac30667ee8dad75286dee"} Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.144155 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.151269 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.151312 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.151335 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.153102 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119"} Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.153151 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115"} Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.153166 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85"} Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.153373 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.154310 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.154355 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.154372 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.356717 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 15:31:48 crc kubenswrapper[4894]: I1209 15:31:48.361507 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.159871 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd"} Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.159919 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.161035 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.161065 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.161073 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.163244 4894 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="bfabeb0a86f0cfb4cc7d7951f7af9579ca1ffcf2287fc55c6da17b388f2bff76" exitCode=0 Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.163294 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"bfabeb0a86f0cfb4cc7d7951f7af9579ca1ffcf2287fc55c6da17b388f2bff76"} Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.163341 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.163422 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.163503 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.165702 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.165744 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.165756 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.165944 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.165975 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.165997 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.166560 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.166596 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.166609 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.280501 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.281882 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.281933 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.281949 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.281977 4894 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 15:31:49 crc kubenswrapper[4894]: I1209 15:31:49.758730 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 15:31:50 crc kubenswrapper[4894]: I1209 15:31:50.170972 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"887ed9d79d00701d91883050354379c6e185c4ae0ff42b90ac0ad68d9cc92ae3"} Dec 09 15:31:50 crc kubenswrapper[4894]: I1209 15:31:50.171027 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"074f035260bc06beb10f4eb1cc71e1b09b6c9c2c1648355bf1d16f60817d738b"} Dec 09 15:31:50 crc kubenswrapper[4894]: I1209 15:31:50.171083 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"a2196a8cecbca116f74e967d215394b158bfc96cec784662bb8526845b3a29ad"} Dec 09 15:31:50 crc kubenswrapper[4894]: I1209 15:31:50.171091 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 15:31:50 crc kubenswrapper[4894]: I1209 15:31:50.171139 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:50 crc kubenswrapper[4894]: I1209 15:31:50.171193 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:50 crc kubenswrapper[4894]: I1209 15:31:50.171096 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ffddeb6eccff302f8ae25a84a897eea7b08c6b54ec7ffb2f073d89e1754b4da7"} Dec 09 15:31:50 crc kubenswrapper[4894]: I1209 15:31:50.172258 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:50 crc kubenswrapper[4894]: I1209 15:31:50.172279 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:50 crc kubenswrapper[4894]: I1209 15:31:50.172292 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:50 crc kubenswrapper[4894]: I1209 15:31:50.172594 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:50 crc kubenswrapper[4894]: I1209 15:31:50.172659 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:50 crc kubenswrapper[4894]: I1209 15:31:50.172676 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:50 crc kubenswrapper[4894]: I1209 15:31:50.249151 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:31:51 crc kubenswrapper[4894]: I1209 15:31:51.175280 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"c90e389a9607bcde09abf18928726743606d9f3c3f182c57d181f95409f0511d"} Dec 09 15:31:51 crc kubenswrapper[4894]: I1209 15:31:51.175325 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 15:31:51 crc kubenswrapper[4894]: I1209 15:31:51.175356 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:51 crc kubenswrapper[4894]: I1209 15:31:51.175388 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:51 crc kubenswrapper[4894]: I1209 15:31:51.175399 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:51 crc kubenswrapper[4894]: I1209 15:31:51.176272 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:51 crc kubenswrapper[4894]: I1209 15:31:51.176304 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:51 crc kubenswrapper[4894]: I1209 15:31:51.176318 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:51 crc kubenswrapper[4894]: I1209 15:31:51.176331 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:51 crc kubenswrapper[4894]: I1209 15:31:51.176347 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:51 crc kubenswrapper[4894]: I1209 15:31:51.176358 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:51 crc kubenswrapper[4894]: I1209 15:31:51.176372 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:51 crc kubenswrapper[4894]: I1209 15:31:51.176395 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:51 crc kubenswrapper[4894]: I1209 15:31:51.176406 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:51 crc kubenswrapper[4894]: I1209 15:31:51.735397 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:31:52 crc kubenswrapper[4894]: I1209 15:31:52.178172 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 15:31:52 crc kubenswrapper[4894]: I1209 15:31:52.178234 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:52 crc kubenswrapper[4894]: I1209 15:31:52.178262 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:52 crc kubenswrapper[4894]: I1209 15:31:52.179373 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:52 crc kubenswrapper[4894]: I1209 15:31:52.179432 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:52 crc kubenswrapper[4894]: I1209 15:31:52.179453 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:52 crc kubenswrapper[4894]: I1209 15:31:52.179487 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:52 crc kubenswrapper[4894]: I1209 15:31:52.179510 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:52 crc kubenswrapper[4894]: I1209 15:31:52.179521 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:52 crc kubenswrapper[4894]: I1209 15:31:52.385844 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Dec 09 15:31:53 crc kubenswrapper[4894]: I1209 15:31:53.181193 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:53 crc kubenswrapper[4894]: I1209 15:31:53.182400 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:53 crc kubenswrapper[4894]: I1209 15:31:53.182466 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:53 crc kubenswrapper[4894]: I1209 15:31:53.182482 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:53 crc kubenswrapper[4894]: I1209 15:31:53.819586 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Dec 09 15:31:53 crc kubenswrapper[4894]: I1209 15:31:53.957556 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 15:31:53 crc kubenswrapper[4894]: I1209 15:31:53.957852 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:53 crc kubenswrapper[4894]: I1209 15:31:53.959342 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:53 crc kubenswrapper[4894]: I1209 15:31:53.959390 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:53 crc kubenswrapper[4894]: I1209 15:31:53.959414 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:54 crc kubenswrapper[4894]: I1209 15:31:54.183000 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:54 crc kubenswrapper[4894]: I1209 15:31:54.183752 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:54 crc kubenswrapper[4894]: I1209 15:31:54.183783 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:54 crc kubenswrapper[4894]: I1209 15:31:54.183792 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:54 crc kubenswrapper[4894]: I1209 15:31:54.795508 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:31:54 crc kubenswrapper[4894]: I1209 15:31:54.795810 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:54 crc kubenswrapper[4894]: I1209 15:31:54.797099 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:54 crc kubenswrapper[4894]: I1209 15:31:54.797168 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:54 crc kubenswrapper[4894]: I1209 15:31:54.797186 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:55 crc kubenswrapper[4894]: I1209 15:31:55.503139 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 15:31:55 crc kubenswrapper[4894]: I1209 15:31:55.503298 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:55 crc kubenswrapper[4894]: I1209 15:31:55.504605 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:55 crc kubenswrapper[4894]: I1209 15:31:55.504702 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:55 crc kubenswrapper[4894]: I1209 15:31:55.504722 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:56 crc kubenswrapper[4894]: E1209 15:31:56.175411 4894 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Dec 09 15:31:56 crc kubenswrapper[4894]: I1209 15:31:56.206526 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 15:31:56 crc kubenswrapper[4894]: I1209 15:31:56.206739 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:56 crc kubenswrapper[4894]: I1209 15:31:56.207676 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:56 crc kubenswrapper[4894]: I1209 15:31:56.207712 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:56 crc kubenswrapper[4894]: I1209 15:31:56.207726 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:31:56 crc kubenswrapper[4894]: I1209 15:31:56.958156 4894 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 15:31:56 crc kubenswrapper[4894]: I1209 15:31:56.958544 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 15:31:58 crc kubenswrapper[4894]: I1209 15:31:58.751458 4894 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 09 15:31:58 crc kubenswrapper[4894]: I1209 15:31:58.752144 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 09 15:31:58 crc kubenswrapper[4894]: I1209 15:31:58.760090 4894 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Dec 09 15:31:58 crc kubenswrapper[4894]: I1209 15:31:58.760157 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Dec 09 15:31:59 crc kubenswrapper[4894]: I1209 15:31:59.763002 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 15:31:59 crc kubenswrapper[4894]: I1209 15:31:59.763254 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:31:59 crc kubenswrapper[4894]: I1209 15:31:59.765330 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:31:59 crc kubenswrapper[4894]: I1209 15:31:59.765375 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:31:59 crc kubenswrapper[4894]: I1209 15:31:59.765390 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:01 crc kubenswrapper[4894]: I1209 15:32:01.743302 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:32:01 crc kubenswrapper[4894]: I1209 15:32:01.743524 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:32:01 crc kubenswrapper[4894]: I1209 15:32:01.744966 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:01 crc kubenswrapper[4894]: I1209 15:32:01.745022 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:01 crc kubenswrapper[4894]: I1209 15:32:01.745037 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:01 crc kubenswrapper[4894]: I1209 15:32:01.748223 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:32:02 crc kubenswrapper[4894]: I1209 15:32:02.202133 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:32:02 crc kubenswrapper[4894]: I1209 15:32:02.203436 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:02 crc kubenswrapper[4894]: I1209 15:32:02.203484 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:02 crc kubenswrapper[4894]: I1209 15:32:02.203505 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:02 crc kubenswrapper[4894]: I1209 15:32:02.407959 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Dec 09 15:32:02 crc kubenswrapper[4894]: I1209 15:32:02.408114 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:32:02 crc kubenswrapper[4894]: I1209 15:32:02.412224 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:02 crc kubenswrapper[4894]: I1209 15:32:02.412283 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:02 crc kubenswrapper[4894]: I1209 15:32:02.412294 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:02 crc kubenswrapper[4894]: I1209 15:32:02.426557 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Dec 09 15:32:03 crc kubenswrapper[4894]: I1209 15:32:03.204110 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:32:03 crc kubenswrapper[4894]: I1209 15:32:03.205420 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:03 crc kubenswrapper[4894]: I1209 15:32:03.205467 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:03 crc kubenswrapper[4894]: I1209 15:32:03.205483 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:03 crc kubenswrapper[4894]: I1209 15:32:03.752993 4894 trace.go:236] Trace[1836824719]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Dec-2025 15:31:49.256) (total time: 14496ms): Dec 09 15:32:03 crc kubenswrapper[4894]: Trace[1836824719]: ---"Objects listed" error: 14496ms (15:32:03.752) Dec 09 15:32:03 crc kubenswrapper[4894]: Trace[1836824719]: [14.496533527s] [14.496533527s] END Dec 09 15:32:03 crc kubenswrapper[4894]: I1209 15:32:03.753027 4894 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 09 15:32:04 crc kubenswrapper[4894]: E1209 15:32:04.080619 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="3.2s" Dec 09 15:32:04 crc kubenswrapper[4894]: E1209 15:32:04.086089 4894 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Dec 09 15:32:04 crc kubenswrapper[4894]: I1209 15:32:04.086299 4894 trace.go:236] Trace[776363883]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Dec-2025 15:31:49.640) (total time: 14445ms): Dec 09 15:32:04 crc kubenswrapper[4894]: Trace[776363883]: ---"Objects listed" error: 14445ms (15:32:04.086) Dec 09 15:32:04 crc kubenswrapper[4894]: Trace[776363883]: [14.445532919s] [14.445532919s] END Dec 09 15:32:04 crc kubenswrapper[4894]: I1209 15:32:04.086320 4894 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 09 15:32:04 crc kubenswrapper[4894]: I1209 15:32:04.086906 4894 trace.go:236] Trace[197198668]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Dec-2025 15:31:49.916) (total time: 14170ms): Dec 09 15:32:04 crc kubenswrapper[4894]: Trace[197198668]: ---"Objects listed" error: 14170ms (15:32:04.086) Dec 09 15:32:04 crc kubenswrapper[4894]: Trace[197198668]: [14.170473258s] [14.170473258s] END Dec 09 15:32:04 crc kubenswrapper[4894]: I1209 15:32:04.086933 4894 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 09 15:32:04 crc kubenswrapper[4894]: I1209 15:32:04.087262 4894 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Dec 09 15:32:04 crc kubenswrapper[4894]: I1209 15:32:04.087374 4894 trace.go:236] Trace[1596218493]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (09-Dec-2025 15:31:49.981) (total time: 14105ms): Dec 09 15:32:04 crc kubenswrapper[4894]: Trace[1596218493]: ---"Objects listed" error: 14105ms (15:32:04.087) Dec 09 15:32:04 crc kubenswrapper[4894]: Trace[1596218493]: [14.105772762s] [14.105772762s] END Dec 09 15:32:04 crc kubenswrapper[4894]: I1209 15:32:04.087399 4894 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 09 15:32:04 crc kubenswrapper[4894]: I1209 15:32:04.146455 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 15:32:04 crc kubenswrapper[4894]: I1209 15:32:04.167335 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 15:32:04 crc kubenswrapper[4894]: E1209 15:32:04.210522 4894 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 15:32:04 crc kubenswrapper[4894]: I1209 15:32:04.241961 4894 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:45530->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 09 15:32:04 crc kubenswrapper[4894]: I1209 15:32:04.242017 4894 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:45534->192.168.126.11:17697: read: connection reset by peer" start-of-body= Dec 09 15:32:04 crc kubenswrapper[4894]: I1209 15:32:04.242074 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:45530->192.168.126.11:17697: read: connection reset by peer" Dec 09 15:32:04 crc kubenswrapper[4894]: I1209 15:32:04.242095 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:45534->192.168.126.11:17697: read: connection reset by peer" Dec 09 15:32:04 crc kubenswrapper[4894]: I1209 15:32:04.242710 4894 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 09 15:32:04 crc kubenswrapper[4894]: I1209 15:32:04.242766 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 09 15:32:04 crc kubenswrapper[4894]: I1209 15:32:04.796558 4894 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Dec 09 15:32:04 crc kubenswrapper[4894]: I1209 15:32:04.796625 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.081313 4894 apiserver.go:52] "Watching apiserver" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.086494 4894 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.086823 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-qt56s","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-operator/iptables-alerter-4ln5h","openshift-dns/node-resolver-mw747","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g"] Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.087142 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.087247 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.087307 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.087320 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.087365 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.087495 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.087885 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.088233 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.088275 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.088461 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-mw747" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.088463 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.091959 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.092601 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.093129 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.094323 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.095581 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.108083 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.108599 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.108821 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.108992 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.109184 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.109202 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.109336 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.109379 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.109473 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.109475 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.109545 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.109708 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.136504 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.148198 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.160966 4894 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.163291 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.177053 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.187240 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.193989 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194036 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194081 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194105 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194230 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194254 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194273 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194406 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194427 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194431 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194474 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194555 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194566 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194620 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194629 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194680 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194710 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194732 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194752 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194770 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194790 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194814 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194835 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194853 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194872 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194923 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194947 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.194971 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195042 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195068 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195091 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195115 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195144 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195167 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195190 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195210 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195232 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195254 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195277 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195300 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195326 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195366 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195392 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195413 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195433 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195452 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195482 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195503 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195521 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195540 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195560 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195580 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195630 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195670 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195690 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195710 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195730 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195752 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195782 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195800 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195840 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195867 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195888 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195907 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195926 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195948 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195971 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195991 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196030 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196054 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196077 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196098 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196123 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196145 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196168 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196193 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196217 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196239 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196262 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196283 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196307 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196331 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196378 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196400 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196424 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196445 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196471 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196492 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196515 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196538 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196566 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196590 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196612 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196648 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196671 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196692 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196712 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196734 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196756 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196778 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196800 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196821 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196843 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196865 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196886 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196907 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196928 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196953 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196976 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196997 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197022 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197044 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197064 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197086 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197108 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197130 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197153 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197385 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197408 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197435 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197469 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197500 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197532 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197566 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197591 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197616 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197757 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197787 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197812 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197838 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197863 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197887 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197921 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197947 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197972 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197995 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198018 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198040 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198063 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198085 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198111 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198137 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198164 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198189 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198220 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198247 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198270 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198296 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198327 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198352 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198373 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198398 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198441 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198467 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198489 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198513 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198539 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198564 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198590 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198617 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198668 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198695 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198729 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198757 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198784 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198807 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198831 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198854 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198882 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198913 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198940 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198967 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198989 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199045 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199070 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199178 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199206 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199231 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199257 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199283 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199311 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199335 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199360 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199582 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199607 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199631 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199674 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199701 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199727 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199751 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199775 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199800 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199828 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199851 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199877 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199908 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199934 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199958 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199984 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200039 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200070 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200099 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/4b263411-430a-4950-8e4d-b0f305874889-hosts-file\") pod \"node-resolver-mw747\" (UID: \"4b263411-430a-4950-8e4d-b0f305874889\") " pod="openshift-dns/node-resolver-mw747" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200126 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200149 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-multus-conf-dir\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200170 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/97d690f0-ac29-46c9-ad4a-b92bd0a30772-multus-daemon-config\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200201 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200228 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200253 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200278 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-host-run-netns\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200303 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czcbk\" (UniqueName: \"kubernetes.io/projected/97d690f0-ac29-46c9-ad4a-b92bd0a30772-kube-api-access-czcbk\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200331 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200357 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200385 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-cnibin\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200412 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200438 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200461 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/97d690f0-ac29-46c9-ad4a-b92bd0a30772-cni-binary-copy\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200482 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-host-var-lib-kubelet\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200504 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200524 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-os-release\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200548 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200575 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-host-run-k8s-cni-cncf-io\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200609 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-host-var-lib-cni-multus\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200683 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-hostroot\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200720 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-etc-kubernetes\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200751 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200778 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.201185 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.206494 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-multus-cni-dir\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.206663 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-host-run-multus-certs\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.206776 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f4wws\" (UniqueName: \"kubernetes.io/projected/4b263411-430a-4950-8e4d-b0f305874889-kube-api-access-f4wws\") pod \"node-resolver-mw747\" (UID: \"4b263411-430a-4950-8e4d-b0f305874889\") " pod="openshift-dns/node-resolver-mw747" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.206882 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-system-cni-dir\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.206972 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-multus-socket-dir-parent\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.207094 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-host-var-lib-cni-bin\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.207244 4894 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.207338 4894 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.207424 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.207513 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195090 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.209429 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195197 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195209 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195270 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195295 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195360 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195753 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195767 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.195988 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196001 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196143 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196276 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196433 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196437 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196497 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196571 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196792 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.196927 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.197951 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198272 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198372 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198505 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198570 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198589 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198689 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198828 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198855 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.198921 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199064 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199343 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199367 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199776 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199857 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.199877 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200025 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200050 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200087 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200184 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200728 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200756 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200759 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200814 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200831 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.200878 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.201071 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.201137 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.201167 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.201433 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.201494 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.201612 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.201792 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.201951 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.201960 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.202043 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.202250 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.202275 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.202302 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.202584 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.202629 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.202700 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.202923 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.203255 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.202686 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.203496 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.203563 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.203118 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.205687 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.206050 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.206441 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.207239 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.207515 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.207943 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.207980 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.208239 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.208294 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.208448 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.208415 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.208591 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.208825 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.209119 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.209174 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.209412 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.210056 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.210116 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.210943 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.211169 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.211431 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.211954 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.211986 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.212244 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.212253 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.212320 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.212598 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.213103 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.213132 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.213216 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.213542 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.213807 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.214046 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.214293 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.214473 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.214733 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.214933 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.214948 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.215187 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.215463 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.215742 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.215828 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.216000 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.216162 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.216572 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.216821 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.217165 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.217434 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.217454 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.217470 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.217577 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.217656 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.217711 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.217839 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.217941 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.217997 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.218010 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.218067 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.218676 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.218809 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.219027 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.219040 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.219175 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.219385 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.219685 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.219763 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.219885 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.219962 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.220111 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:32:05.720090521 +0000 UTC m=+20.039301280 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.220200 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.220211 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.220394 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.220619 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.220814 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.220884 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.220926 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:05.720915622 +0000 UTC m=+20.040126291 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.221860 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.221941 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.222130 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:05.722077963 +0000 UTC m=+20.041288712 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.222228 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.222320 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.222620 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.222664 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.223324 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.226442 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.226914 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.227480 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.228162 4894 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.228514 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.228872 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.228961 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.229133 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.229157 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.229586 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.224600 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.224724 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.225153 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.226177 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.229722 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.229938 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.229957 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.229993 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.230009 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.230223 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.230411 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.230437 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.230709 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.230949 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.233854 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.234375 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.225815 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.234863 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.231839 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.232041 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.232697 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.233211 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.236546 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.236884 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.237567 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.237853 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.238117 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.240480 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.242535 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.245971 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.246167 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.246270 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.246377 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.246491 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:05.746383203 +0000 UTC m=+20.065593942 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.246763 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.246931 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.247090 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.247983 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.256623 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.257156 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.257189 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.257200 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.257252 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:05.757236109 +0000 UTC m=+20.076446778 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.259241 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.260625 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.260874 4894 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd" exitCode=255 Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.261166 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.260939 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd"} Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.260874 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.265054 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.265527 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.266202 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.266359 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.266804 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.276455 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.277708 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.286598 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.289973 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.292506 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.306006 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.306418 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-54wfs"] Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.309978 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-rj7cb"] Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.310437 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/97d690f0-ac29-46c9-ad4a-b92bd0a30772-cni-binary-copy\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.310513 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-host-var-lib-kubelet\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.310549 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.310602 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-os-release\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.310665 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-host-run-k8s-cni-cncf-io\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.310695 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-host-var-lib-cni-multus\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.310731 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-hostroot\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.310761 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-etc-kubernetes\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.310792 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-multus-cni-dir\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.310820 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-host-run-multus-certs\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.310848 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f4wws\" (UniqueName: \"kubernetes.io/projected/4b263411-430a-4950-8e4d-b0f305874889-kube-api-access-f4wws\") pod \"node-resolver-mw747\" (UID: \"4b263411-430a-4950-8e4d-b0f305874889\") " pod="openshift-dns/node-resolver-mw747" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.310883 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-system-cni-dir\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.310917 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-multus-socket-dir-parent\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.310960 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-host-var-lib-cni-bin\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.310988 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/4b263411-430a-4950-8e4d-b0f305874889-hosts-file\") pod \"node-resolver-mw747\" (UID: \"4b263411-430a-4950-8e4d-b0f305874889\") " pod="openshift-dns/node-resolver-mw747" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.311021 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-multus-conf-dir\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.311045 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/97d690f0-ac29-46c9-ad4a-b92bd0a30772-multus-daemon-config\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.311221 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-host-run-netns\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.311250 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czcbk\" (UniqueName: \"kubernetes.io/projected/97d690f0-ac29-46c9-ad4a-b92bd0a30772-kube-api-access-czcbk\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.311279 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.311281 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.311352 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-cnibin\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.311716 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.311737 4894 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.311754 4894 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.311767 4894 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.311779 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.311791 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.311806 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.311818 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.311829 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.311840 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.311855 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.311871 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.311885 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.311896 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.312067 4894 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.312084 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.312097 4894 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.312113 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.312125 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.312274 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-cnibin\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.312760 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-host-var-lib-kubelet\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.312963 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.313024 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-os-release\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.313056 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-host-run-k8s-cni-cncf-io\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.313094 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-host-var-lib-cni-multus\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.313121 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-hostroot\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.313153 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-etc-kubernetes\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.313211 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-multus-cni-dir\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.313266 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-host-run-multus-certs\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.313525 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-system-cni-dir\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.313574 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-multus-socket-dir-parent\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.313606 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-host-var-lib-cni-bin\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.313659 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/4b263411-430a-4950-8e4d-b0f305874889-hosts-file\") pod \"node-resolver-mw747\" (UID: \"4b263411-430a-4950-8e4d-b0f305874889\") " pod="openshift-dns/node-resolver-mw747" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.313686 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-multus-conf-dir\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.312125 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314129 4894 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314145 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314157 4894 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314173 4894 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314185 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314198 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314210 4894 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314228 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314278 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314291 4894 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314305 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314345 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314359 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314372 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314398 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314416 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314428 4894 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314441 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314458 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314472 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314484 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314497 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314513 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314525 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314537 4894 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314579 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314596 4894 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314609 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314621 4894 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314655 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314669 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314682 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314695 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314712 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314726 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314739 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314753 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314770 4894 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314783 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314812 4894 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314826 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314843 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314855 4894 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314882 4894 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314898 4894 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314911 4894 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314943 4894 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314977 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.314995 4894 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315008 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315033 4894 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315045 4894 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315061 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315073 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315102 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315114 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315128 4894 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315137 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315146 4894 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315157 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315316 4894 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315347 4894 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315356 4894 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315367 4894 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315376 4894 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315385 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315394 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315406 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315437 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315446 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315458 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315466 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315474 4894 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315597 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/97d690f0-ac29-46c9-ad4a-b92bd0a30772-host-run-netns\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315796 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315865 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315888 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315927 4894 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315938 4894 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315948 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315957 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315968 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315976 4894 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315985 4894 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.315995 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316006 4894 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316017 4894 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316026 4894 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316037 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316045 4894 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316054 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316063 4894 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316074 4894 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316084 4894 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316092 4894 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316102 4894 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316114 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316123 4894 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316131 4894 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316142 4894 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316151 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316161 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316188 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316325 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316347 4894 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316445 4894 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316484 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316504 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316518 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316531 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316551 4894 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316565 4894 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316580 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316711 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316732 4894 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316746 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316784 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316802 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316815 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316827 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316846 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316858 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316873 4894 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316912 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316968 4894 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.316998 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317021 4894 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317033 4894 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317044 4894 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317059 4894 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317068 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317072 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317098 4894 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317147 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317163 4894 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317200 4894 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317212 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317226 4894 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317238 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317249 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317260 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317273 4894 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317304 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317321 4894 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317332 4894 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317346 4894 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317358 4894 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317368 4894 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317378 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317392 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317402 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317413 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.317426 4894 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.318812 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.319499 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.319928 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.320000 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322024 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322127 4894 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322143 4894 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322154 4894 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322164 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322178 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322189 4894 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322199 4894 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322208 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322217 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322226 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322235 4894 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322244 4894 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322252 4894 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322261 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322271 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322280 4894 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322290 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322300 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322309 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322318 4894 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322326 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322373 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322412 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/97d690f0-ac29-46c9-ad4a-b92bd0a30772-multus-daemon-config\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.322432 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/97d690f0-ac29-46c9-ad4a-b92bd0a30772-cni-binary-copy\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.324446 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.333676 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czcbk\" (UniqueName: \"kubernetes.io/projected/97d690f0-ac29-46c9-ad4a-b92bd0a30772-kube-api-access-czcbk\") pod \"multus-qt56s\" (UID: \"97d690f0-ac29-46c9-ad4a-b92bd0a30772\") " pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.338961 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.339449 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f4wws\" (UniqueName: \"kubernetes.io/projected/4b263411-430a-4950-8e4d-b0f305874889-kube-api-access-f4wws\") pod \"node-resolver-mw747\" (UID: \"4b263411-430a-4950-8e4d-b0f305874889\") " pod="openshift-dns/node-resolver-mw747" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.339471 4894 scope.go:117] "RemoveContainer" containerID="5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.341225 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.357255 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.368457 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.388080 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.401198 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.409381 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.414290 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.417771 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.423756 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phrks\" (UniqueName: \"kubernetes.io/projected/cfa97fcb-7e7c-4202-a2d0-d283fa49b112-kube-api-access-phrks\") pod \"machine-config-daemon-rj7cb\" (UID: \"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\") " pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.423816 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/05e6de1b-a262-4d89-ad3b-902a436d44cc-system-cni-dir\") pod \"multus-additional-cni-plugins-54wfs\" (UID: \"05e6de1b-a262-4d89-ad3b-902a436d44cc\") " pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.423902 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/cfa97fcb-7e7c-4202-a2d0-d283fa49b112-rootfs\") pod \"machine-config-daemon-rj7cb\" (UID: \"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\") " pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.423934 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/05e6de1b-a262-4d89-ad3b-902a436d44cc-cnibin\") pod \"multus-additional-cni-plugins-54wfs\" (UID: \"05e6de1b-a262-4d89-ad3b-902a436d44cc\") " pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.423992 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/05e6de1b-a262-4d89-ad3b-902a436d44cc-cni-binary-copy\") pod \"multus-additional-cni-plugins-54wfs\" (UID: \"05e6de1b-a262-4d89-ad3b-902a436d44cc\") " pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.424023 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/05e6de1b-a262-4d89-ad3b-902a436d44cc-tuning-conf-dir\") pod \"multus-additional-cni-plugins-54wfs\" (UID: \"05e6de1b-a262-4d89-ad3b-902a436d44cc\") " pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.424072 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cfa97fcb-7e7c-4202-a2d0-d283fa49b112-mcd-auth-proxy-config\") pod \"machine-config-daemon-rj7cb\" (UID: \"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\") " pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.424095 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/05e6de1b-a262-4d89-ad3b-902a436d44cc-os-release\") pod \"multus-additional-cni-plugins-54wfs\" (UID: \"05e6de1b-a262-4d89-ad3b-902a436d44cc\") " pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.424123 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cfa97fcb-7e7c-4202-a2d0-d283fa49b112-proxy-tls\") pod \"machine-config-daemon-rj7cb\" (UID: \"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\") " pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.424143 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/05e6de1b-a262-4d89-ad3b-902a436d44cc-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-54wfs\" (UID: \"05e6de1b-a262-4d89-ad3b-902a436d44cc\") " pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.424161 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jm69\" (UniqueName: \"kubernetes.io/projected/05e6de1b-a262-4d89-ad3b-902a436d44cc-kube-api-access-5jm69\") pod \"multus-additional-cni-plugins-54wfs\" (UID: \"05e6de1b-a262-4d89-ad3b-902a436d44cc\") " pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: W1209 15:32:05.424728 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-7b5c8ba87e8736ce6621fda74f2f7e8c50b3295678597f03ce7476eb30ca935f WatchSource:0}: Error finding container 7b5c8ba87e8736ce6621fda74f2f7e8c50b3295678597f03ce7476eb30ca935f: Status 404 returned error can't find the container with id 7b5c8ba87e8736ce6621fda74f2f7e8c50b3295678597f03ce7476eb30ca935f Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.425796 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.426165 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Dec 09 15:32:05 crc kubenswrapper[4894]: W1209 15:32:05.430218 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-9c931e9583b41567a28eb7972e3290706a7202278d141da290271d8f82e27ac0 WatchSource:0}: Error finding container 9c931e9583b41567a28eb7972e3290706a7202278d141da290271d8f82e27ac0: Status 404 returned error can't find the container with id 9c931e9583b41567a28eb7972e3290706a7202278d141da290271d8f82e27ac0 Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.437377 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.438520 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-qt56s" Dec 09 15:32:05 crc kubenswrapper[4894]: W1209 15:32:05.439723 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-80b0bf3638f2d506700035b579e948762a3af7e7eeb1e0953723e936d2ec3bdb WatchSource:0}: Error finding container 80b0bf3638f2d506700035b579e948762a3af7e7eeb1e0953723e936d2ec3bdb: Status 404 returned error can't find the container with id 80b0bf3638f2d506700035b579e948762a3af7e7eeb1e0953723e936d2ec3bdb Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.445592 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-mw747" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.449709 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.465118 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.480961 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: W1209 15:32:05.491006 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97d690f0_ac29_46c9_ad4a_b92bd0a30772.slice/crio-db19edab4ddff3eb595a98db46b706cd5c87f042e54ec29fdd45bc6f3058b79a WatchSource:0}: Error finding container db19edab4ddff3eb595a98db46b706cd5c87f042e54ec29fdd45bc6f3058b79a: Status 404 returned error can't find the container with id db19edab4ddff3eb595a98db46b706cd5c87f042e54ec29fdd45bc6f3058b79a Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.496087 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.522904 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.526963 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/05e6de1b-a262-4d89-ad3b-902a436d44cc-system-cni-dir\") pod \"multus-additional-cni-plugins-54wfs\" (UID: \"05e6de1b-a262-4d89-ad3b-902a436d44cc\") " pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.527001 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/cfa97fcb-7e7c-4202-a2d0-d283fa49b112-rootfs\") pod \"machine-config-daemon-rj7cb\" (UID: \"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\") " pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.527021 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/05e6de1b-a262-4d89-ad3b-902a436d44cc-cnibin\") pod \"multus-additional-cni-plugins-54wfs\" (UID: \"05e6de1b-a262-4d89-ad3b-902a436d44cc\") " pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.527043 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/05e6de1b-a262-4d89-ad3b-902a436d44cc-cni-binary-copy\") pod \"multus-additional-cni-plugins-54wfs\" (UID: \"05e6de1b-a262-4d89-ad3b-902a436d44cc\") " pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.527060 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/05e6de1b-a262-4d89-ad3b-902a436d44cc-tuning-conf-dir\") pod \"multus-additional-cni-plugins-54wfs\" (UID: \"05e6de1b-a262-4d89-ad3b-902a436d44cc\") " pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.527084 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cfa97fcb-7e7c-4202-a2d0-d283fa49b112-mcd-auth-proxy-config\") pod \"machine-config-daemon-rj7cb\" (UID: \"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\") " pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.527098 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/05e6de1b-a262-4d89-ad3b-902a436d44cc-os-release\") pod \"multus-additional-cni-plugins-54wfs\" (UID: \"05e6de1b-a262-4d89-ad3b-902a436d44cc\") " pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.527113 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/05e6de1b-a262-4d89-ad3b-902a436d44cc-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-54wfs\" (UID: \"05e6de1b-a262-4d89-ad3b-902a436d44cc\") " pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.527130 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jm69\" (UniqueName: \"kubernetes.io/projected/05e6de1b-a262-4d89-ad3b-902a436d44cc-kube-api-access-5jm69\") pod \"multus-additional-cni-plugins-54wfs\" (UID: \"05e6de1b-a262-4d89-ad3b-902a436d44cc\") " pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.527170 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cfa97fcb-7e7c-4202-a2d0-d283fa49b112-proxy-tls\") pod \"machine-config-daemon-rj7cb\" (UID: \"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\") " pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.527187 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phrks\" (UniqueName: \"kubernetes.io/projected/cfa97fcb-7e7c-4202-a2d0-d283fa49b112-kube-api-access-phrks\") pod \"machine-config-daemon-rj7cb\" (UID: \"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\") " pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.527461 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/05e6de1b-a262-4d89-ad3b-902a436d44cc-system-cni-dir\") pod \"multus-additional-cni-plugins-54wfs\" (UID: \"05e6de1b-a262-4d89-ad3b-902a436d44cc\") " pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.527494 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/cfa97fcb-7e7c-4202-a2d0-d283fa49b112-rootfs\") pod \"machine-config-daemon-rj7cb\" (UID: \"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\") " pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.527514 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/05e6de1b-a262-4d89-ad3b-902a436d44cc-cnibin\") pod \"multus-additional-cni-plugins-54wfs\" (UID: \"05e6de1b-a262-4d89-ad3b-902a436d44cc\") " pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.527945 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/05e6de1b-a262-4d89-ad3b-902a436d44cc-os-release\") pod \"multus-additional-cni-plugins-54wfs\" (UID: \"05e6de1b-a262-4d89-ad3b-902a436d44cc\") " pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.528086 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/05e6de1b-a262-4d89-ad3b-902a436d44cc-cni-binary-copy\") pod \"multus-additional-cni-plugins-54wfs\" (UID: \"05e6de1b-a262-4d89-ad3b-902a436d44cc\") " pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.528341 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/05e6de1b-a262-4d89-ad3b-902a436d44cc-tuning-conf-dir\") pod \"multus-additional-cni-plugins-54wfs\" (UID: \"05e6de1b-a262-4d89-ad3b-902a436d44cc\") " pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.528449 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/05e6de1b-a262-4d89-ad3b-902a436d44cc-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-54wfs\" (UID: \"05e6de1b-a262-4d89-ad3b-902a436d44cc\") " pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.532741 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/cfa97fcb-7e7c-4202-a2d0-d283fa49b112-mcd-auth-proxy-config\") pod \"machine-config-daemon-rj7cb\" (UID: \"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\") " pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.537101 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/cfa97fcb-7e7c-4202-a2d0-d283fa49b112-proxy-tls\") pod \"machine-config-daemon-rj7cb\" (UID: \"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\") " pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.537924 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.551425 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.556792 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jm69\" (UniqueName: \"kubernetes.io/projected/05e6de1b-a262-4d89-ad3b-902a436d44cc-kube-api-access-5jm69\") pod \"multus-additional-cni-plugins-54wfs\" (UID: \"05e6de1b-a262-4d89-ad3b-902a436d44cc\") " pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.564016 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phrks\" (UniqueName: \"kubernetes.io/projected/cfa97fcb-7e7c-4202-a2d0-d283fa49b112-kube-api-access-phrks\") pod \"machine-config-daemon-rj7cb\" (UID: \"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\") " pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.567630 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.584788 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.596400 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.633622 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.650836 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-54wfs" Dec 09 15:32:05 crc kubenswrapper[4894]: W1209 15:32:05.651987 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfa97fcb_7e7c_4202_a2d0_d283fa49b112.slice/crio-8c76af16ef8dd9dad3162c9cc3efe25ef3b19203fc7752a96b46f5939790e7c7 WatchSource:0}: Error finding container 8c76af16ef8dd9dad3162c9cc3efe25ef3b19203fc7752a96b46f5939790e7c7: Status 404 returned error can't find the container with id 8c76af16ef8dd9dad3162c9cc3efe25ef3b19203fc7752a96b46f5939790e7c7 Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.667621 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-s5scr"] Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.671837 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.675666 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.676069 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.676420 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.677471 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.677820 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.678160 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.678916 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.697184 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.707907 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.721756 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.731722 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.731837 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-run-ovn-kubernetes\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.731867 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-run-netns\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.731889 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5l7x\" (UniqueName: \"kubernetes.io/projected/97036c6f-dce3-444b-85e6-8476c16613c9-kube-api-access-d5l7x\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.731925 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-slash\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.731949 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-etc-openvswitch\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.731972 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-log-socket\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.731994 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-var-lib-openvswitch\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.732015 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-node-log\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.732035 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-cni-netd\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.732061 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.732083 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-cni-bin\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.732104 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-run-ovn\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.732129 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/97036c6f-dce3-444b-85e6-8476c16613c9-env-overrides\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.732159 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.732181 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-systemd-units\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.732201 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-run-openvswitch\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.732223 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/97036c6f-dce3-444b-85e6-8476c16613c9-ovnkube-script-lib\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.732245 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-kubelet\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.732269 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-run-systemd\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.732292 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.732319 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/97036c6f-dce3-444b-85e6-8476c16613c9-ovnkube-config\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.732340 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/97036c6f-dce3-444b-85e6-8476c16613c9-ovn-node-metrics-cert\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.732456 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:32:06.732438238 +0000 UTC m=+21.051648907 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.732574 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.732623 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:06.732613744 +0000 UTC m=+21.051824503 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.732784 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.732899 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:06.732880621 +0000 UTC m=+21.052091290 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.737864 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.751182 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.761355 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.771342 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.780245 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.791031 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.802619 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.812163 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.827924 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.832908 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-var-lib-openvswitch\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.832946 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-node-log\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.832969 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-cni-netd\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.832992 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-cni-bin\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833024 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-run-ovn\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833042 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/97036c6f-dce3-444b-85e6-8476c16613c9-env-overrides\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833070 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-systemd-units\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833091 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-run-openvswitch\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833110 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/97036c6f-dce3-444b-85e6-8476c16613c9-ovnkube-script-lib\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833130 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-kubelet\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833147 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-run-systemd\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833168 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833190 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/97036c6f-dce3-444b-85e6-8476c16613c9-ovnkube-config\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833213 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/97036c6f-dce3-444b-85e6-8476c16613c9-ovn-node-metrics-cert\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833238 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833261 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-run-ovn-kubernetes\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833283 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-run-netns\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833303 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5l7x\" (UniqueName: \"kubernetes.io/projected/97036c6f-dce3-444b-85e6-8476c16613c9-kube-api-access-d5l7x\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833328 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833349 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-slash\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833364 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-etc-openvswitch\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833393 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-log-socket\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833453 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-log-socket\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833484 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-var-lib-openvswitch\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833507 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-node-log\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833526 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-cni-netd\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833548 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-cni-bin\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.833575 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-run-ovn\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.834065 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/97036c6f-dce3-444b-85e6-8476c16613c9-env-overrides\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.834146 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.834160 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.834169 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.834202 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:06.834191669 +0000 UTC m=+21.153402338 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.834230 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-slash\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.834253 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-etc-openvswitch\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.834370 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-systemd-units\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.834408 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-run-ovn-kubernetes\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.834440 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-run-openvswitch\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.834450 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-run-systemd\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.834494 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.834505 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.834483 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.834544 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-run-netns\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.834520 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:05 crc kubenswrapper[4894]: E1209 15:32:05.834683 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:06.834660151 +0000 UTC m=+21.153870860 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.834418 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-kubelet\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.834955 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/97036c6f-dce3-444b-85e6-8476c16613c9-ovnkube-script-lib\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.835706 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/97036c6f-dce3-444b-85e6-8476c16613c9-ovnkube-config\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.839381 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/97036c6f-dce3-444b-85e6-8476c16613c9-ovn-node-metrics-cert\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.848074 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Dec 09 15:32:05 crc kubenswrapper[4894]: I1209 15:32:05.853009 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5l7x\" (UniqueName: \"kubernetes.io/projected/97036c6f-dce3-444b-85e6-8476c16613c9-kube-api-access-d5l7x\") pod \"ovnkube-node-s5scr\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.017971 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.105479 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:06 crc kubenswrapper[4894]: E1209 15:32:06.105654 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.109687 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.110684 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: W1209 15:32:06.111230 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod97036c6f_dce3_444b_85e6_8476c16613c9.slice/crio-9cd60c8a7e5113f8c18495faaa522a4339748e2680b99424222e32596d12b619 WatchSource:0}: Error finding container 9cd60c8a7e5113f8c18495faaa522a4339748e2680b99424222e32596d12b619: Status 404 returned error can't find the container with id 9cd60c8a7e5113f8c18495faaa522a4339748e2680b99424222e32596d12b619 Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.111674 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.112704 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.114770 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.115587 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.117244 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.118067 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.119361 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.119886 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.121194 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.121892 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.122375 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.123243 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.124115 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.124710 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.125234 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.126383 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.127289 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.128343 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.128913 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.130003 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.130939 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.132062 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.132529 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.134975 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.135785 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.136225 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.137096 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.138113 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.142561 4894 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.142742 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.144499 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.145604 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.146894 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.148496 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.149557 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.150156 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.151590 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.152425 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.153812 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.154537 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.158206 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.158321 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.159086 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.160184 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.160731 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.161625 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.162384 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.163297 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.163881 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.164358 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.165261 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.165846 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.166792 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.207052 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.258763 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.265199 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-mw747" event={"ID":"4b263411-430a-4950-8e4d-b0f305874889","Type":"ContainerStarted","Data":"4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1"} Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.265282 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-mw747" event={"ID":"4b263411-430a-4950-8e4d-b0f305874889","Type":"ContainerStarted","Data":"49902f119796cc6f6eee3f39ba2b241c6e26f55a9a936916148bf4d639df7635"} Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.266891 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qt56s" event={"ID":"97d690f0-ac29-46c9-ad4a-b92bd0a30772","Type":"ContainerStarted","Data":"51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e"} Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.266925 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qt56s" event={"ID":"97d690f0-ac29-46c9-ad4a-b92bd0a30772","Type":"ContainerStarted","Data":"db19edab4ddff3eb595a98db46b706cd5c87f042e54ec29fdd45bc6f3058b79a"} Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.267967 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerStarted","Data":"993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034"} Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.268016 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerStarted","Data":"9cd60c8a7e5113f8c18495faaa522a4339748e2680b99424222e32596d12b619"} Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.272843 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerStarted","Data":"6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd"} Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.272898 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerStarted","Data":"f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f"} Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.272911 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerStarted","Data":"8c76af16ef8dd9dad3162c9cc3efe25ef3b19203fc7752a96b46f5939790e7c7"} Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.273710 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"80b0bf3638f2d506700035b579e948762a3af7e7eeb1e0953723e936d2ec3bdb"} Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.275244 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791"} Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.275277 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13"} Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.275292 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"9c931e9583b41567a28eb7972e3290706a7202278d141da290271d8f82e27ac0"} Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.275997 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.276840 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.278301 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba"} Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.278627 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.279608 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d"} Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.279668 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"7b5c8ba87e8736ce6621fda74f2f7e8c50b3295678597f03ce7476eb30ca935f"} Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.281678 4894 generic.go:334] "Generic (PLEG): container finished" podID="05e6de1b-a262-4d89-ad3b-902a436d44cc" containerID="3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82" exitCode=0 Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.281803 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" event={"ID":"05e6de1b-a262-4d89-ad3b-902a436d44cc","Type":"ContainerDied","Data":"3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82"} Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.281901 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" event={"ID":"05e6de1b-a262-4d89-ad3b-902a436d44cc","Type":"ContainerStarted","Data":"5cacfb7f6c31d34293bd26c4740d13ee99e8ac1dc9939855bf3d46b53e50273c"} Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.300245 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.315719 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.328240 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.341065 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.356241 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.376526 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.388724 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.401743 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.413843 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.429248 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.451553 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.464426 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.479207 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.492778 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.503618 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.513900 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.551626 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.605089 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.634449 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.672790 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.711765 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.748148 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.748240 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.748266 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:06 crc kubenswrapper[4894]: E1209 15:32:06.748285 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:32:08.748266502 +0000 UTC m=+23.067477171 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:32:06 crc kubenswrapper[4894]: E1209 15:32:06.748354 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 15:32:06 crc kubenswrapper[4894]: E1209 15:32:06.748394 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:08.748386035 +0000 UTC m=+23.067596704 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 15:32:06 crc kubenswrapper[4894]: E1209 15:32:06.748413 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 15:32:06 crc kubenswrapper[4894]: E1209 15:32:06.748507 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:08.748484148 +0000 UTC m=+23.067694877 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.750947 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.849208 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:06 crc kubenswrapper[4894]: I1209 15:32:06.849246 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:06 crc kubenswrapper[4894]: E1209 15:32:06.849383 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 15:32:06 crc kubenswrapper[4894]: E1209 15:32:06.849407 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 15:32:06 crc kubenswrapper[4894]: E1209 15:32:06.849419 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:06 crc kubenswrapper[4894]: E1209 15:32:06.849461 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:08.849447938 +0000 UTC m=+23.168658607 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:06 crc kubenswrapper[4894]: E1209 15:32:06.849382 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 15:32:06 crc kubenswrapper[4894]: E1209 15:32:06.849504 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 15:32:06 crc kubenswrapper[4894]: E1209 15:32:06.849516 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:06 crc kubenswrapper[4894]: E1209 15:32:06.849554 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:08.849544 +0000 UTC m=+23.168754669 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.106017 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.106054 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:07 crc kubenswrapper[4894]: E1209 15:32:07.106153 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:07 crc kubenswrapper[4894]: E1209 15:32:07.106254 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.138033 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-6hlsx"] Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.138433 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-6hlsx" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.140100 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.140477 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.140770 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.141602 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.156389 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.168900 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.180374 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.196182 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.214159 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.239130 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.252919 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxzgx\" (UniqueName: \"kubernetes.io/projected/5f2fb540-7624-49a9-8143-aee00c80ae0e-kube-api-access-cxzgx\") pod \"node-ca-6hlsx\" (UID: \"5f2fb540-7624-49a9-8143-aee00c80ae0e\") " pod="openshift-image-registry/node-ca-6hlsx" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.252996 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f2fb540-7624-49a9-8143-aee00c80ae0e-host\") pod \"node-ca-6hlsx\" (UID: \"5f2fb540-7624-49a9-8143-aee00c80ae0e\") " pod="openshift-image-registry/node-ca-6hlsx" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.253022 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5f2fb540-7624-49a9-8143-aee00c80ae0e-serviceca\") pod \"node-ca-6hlsx\" (UID: \"5f2fb540-7624-49a9-8143-aee00c80ae0e\") " pod="openshift-image-registry/node-ca-6hlsx" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.265646 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.283465 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.285585 4894 generic.go:334] "Generic (PLEG): container finished" podID="97036c6f-dce3-444b-85e6-8476c16613c9" containerID="993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034" exitCode=0 Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.285690 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerDied","Data":"993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034"} Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.286506 4894 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.288087 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.288109 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.288124 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.288133 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" event={"ID":"05e6de1b-a262-4d89-ad3b-902a436d44cc","Type":"ContainerStarted","Data":"ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7"} Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.288232 4894 kubelet_node_status.go:76] "Attempting to register node" node="crc" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.303486 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.304751 4894 kubelet_node_status.go:115] "Node was previously registered" node="crc" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.304974 4894 kubelet_node_status.go:79] "Successfully registered node" node="crc" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.305823 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.305840 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.305848 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.305897 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.305906 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:07Z","lastTransitionTime":"2025-12-09T15:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.316605 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: E1209 15:32:07.320264 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.323449 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.323482 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.323493 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.323506 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.323515 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:07Z","lastTransitionTime":"2025-12-09T15:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.329301 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: E1209 15:32:07.335861 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.338777 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.338819 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.338831 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.338851 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.338863 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:07Z","lastTransitionTime":"2025-12-09T15:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.351688 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: E1209 15:32:07.351772 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.354535 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxzgx\" (UniqueName: \"kubernetes.io/projected/5f2fb540-7624-49a9-8143-aee00c80ae0e-kube-api-access-cxzgx\") pod \"node-ca-6hlsx\" (UID: \"5f2fb540-7624-49a9-8143-aee00c80ae0e\") " pod="openshift-image-registry/node-ca-6hlsx" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.354660 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f2fb540-7624-49a9-8143-aee00c80ae0e-host\") pod \"node-ca-6hlsx\" (UID: \"5f2fb540-7624-49a9-8143-aee00c80ae0e\") " pod="openshift-image-registry/node-ca-6hlsx" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.354684 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5f2fb540-7624-49a9-8143-aee00c80ae0e-serviceca\") pod \"node-ca-6hlsx\" (UID: \"5f2fb540-7624-49a9-8143-aee00c80ae0e\") " pod="openshift-image-registry/node-ca-6hlsx" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.354833 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5f2fb540-7624-49a9-8143-aee00c80ae0e-host\") pod \"node-ca-6hlsx\" (UID: \"5f2fb540-7624-49a9-8143-aee00c80ae0e\") " pod="openshift-image-registry/node-ca-6hlsx" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.357145 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.357189 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.357199 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.357217 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.357229 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:07Z","lastTransitionTime":"2025-12-09T15:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.360374 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5f2fb540-7624-49a9-8143-aee00c80ae0e-serviceca\") pod \"node-ca-6hlsx\" (UID: \"5f2fb540-7624-49a9-8143-aee00c80ae0e\") " pod="openshift-image-registry/node-ca-6hlsx" Dec 09 15:32:07 crc kubenswrapper[4894]: E1209 15:32:07.376499 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.382039 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.382093 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.382106 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.382128 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.382143 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:07Z","lastTransitionTime":"2025-12-09T15:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:07 crc kubenswrapper[4894]: E1209 15:32:07.395370 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: E1209 15:32:07.395964 4894 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.397839 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.397939 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.398027 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.398117 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.398175 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:07Z","lastTransitionTime":"2025-12-09T15:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.403751 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxzgx\" (UniqueName: \"kubernetes.io/projected/5f2fb540-7624-49a9-8143-aee00c80ae0e-kube-api-access-cxzgx\") pod \"node-ca-6hlsx\" (UID: \"5f2fb540-7624-49a9-8143-aee00c80ae0e\") " pod="openshift-image-registry/node-ca-6hlsx" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.414356 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.450988 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.494562 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.499982 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.500024 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.500034 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.500050 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.500061 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:07Z","lastTransitionTime":"2025-12-09T15:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.533356 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.573563 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.602632 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.602695 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.602708 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.602727 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.602741 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:07Z","lastTransitionTime":"2025-12-09T15:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.616913 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.651243 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.659667 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-6hlsx" Dec 09 15:32:07 crc kubenswrapper[4894]: W1209 15:32:07.681316 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5f2fb540_7624_49a9_8143_aee00c80ae0e.slice/crio-ec0b3ae3890eef11afd1107bce49bc0b1bd24054e2fecd3ba4c96d3452a2fc4a WatchSource:0}: Error finding container ec0b3ae3890eef11afd1107bce49bc0b1bd24054e2fecd3ba4c96d3452a2fc4a: Status 404 returned error can't find the container with id ec0b3ae3890eef11afd1107bce49bc0b1bd24054e2fecd3ba4c96d3452a2fc4a Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.694773 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.707123 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.707161 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.707172 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.707188 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.707199 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:07Z","lastTransitionTime":"2025-12-09T15:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.747520 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.774176 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.810455 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.810501 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.810513 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.810531 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.810547 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:07Z","lastTransitionTime":"2025-12-09T15:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.812166 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.853533 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.893722 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.913088 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.913125 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.913135 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.913149 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.913159 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:07Z","lastTransitionTime":"2025-12-09T15:32:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.933314 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:07 crc kubenswrapper[4894]: I1209 15:32:07.973120 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:07Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.013649 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.018977 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.019045 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.019061 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.019082 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.019426 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:08Z","lastTransitionTime":"2025-12-09T15:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.106980 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:08 crc kubenswrapper[4894]: E1209 15:32:08.107139 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.122141 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.122176 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.122186 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.122201 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.122213 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:08Z","lastTransitionTime":"2025-12-09T15:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.224381 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.224419 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.224428 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.224441 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.224451 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:08Z","lastTransitionTime":"2025-12-09T15:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.292765 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-6hlsx" event={"ID":"5f2fb540-7624-49a9-8143-aee00c80ae0e","Type":"ContainerStarted","Data":"b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037"} Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.292815 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-6hlsx" event={"ID":"5f2fb540-7624-49a9-8143-aee00c80ae0e","Type":"ContainerStarted","Data":"ec0b3ae3890eef11afd1107bce49bc0b1bd24054e2fecd3ba4c96d3452a2fc4a"} Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.294247 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4"} Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.297697 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerStarted","Data":"9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249"} Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.297735 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerStarted","Data":"860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102"} Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.297745 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerStarted","Data":"4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed"} Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.297756 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerStarted","Data":"91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8"} Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.297767 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerStarted","Data":"cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90"} Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.297775 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerStarted","Data":"56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8"} Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.299460 4894 generic.go:334] "Generic (PLEG): container finished" podID="05e6de1b-a262-4d89-ad3b-902a436d44cc" containerID="ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7" exitCode=0 Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.299486 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" event={"ID":"05e6de1b-a262-4d89-ad3b-902a436d44cc","Type":"ContainerDied","Data":"ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7"} Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.307621 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.323618 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.326378 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.326422 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.326435 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.326453 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.326467 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:08Z","lastTransitionTime":"2025-12-09T15:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.335776 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.346073 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.361471 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.380453 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.393963 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.406885 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.416480 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.429777 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.429819 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.429827 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.429842 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.429852 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:08Z","lastTransitionTime":"2025-12-09T15:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.431621 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.457494 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.491418 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.531379 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.532270 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.532302 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.532313 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.532330 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.532341 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:08Z","lastTransitionTime":"2025-12-09T15:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.572847 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.612764 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.634874 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.634933 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.634944 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.634960 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.634973 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:08Z","lastTransitionTime":"2025-12-09T15:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.653946 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.694005 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.730299 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.736801 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.736829 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.736838 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.736851 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.736861 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:08Z","lastTransitionTime":"2025-12-09T15:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.768253 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.768363 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.768390 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:08 crc kubenswrapper[4894]: E1209 15:32:08.768422 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:32:12.768391416 +0000 UTC m=+27.087602085 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:32:08 crc kubenswrapper[4894]: E1209 15:32:08.768465 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 15:32:08 crc kubenswrapper[4894]: E1209 15:32:08.768483 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 15:32:08 crc kubenswrapper[4894]: E1209 15:32:08.768530 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:12.768515539 +0000 UTC m=+27.087726278 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 15:32:08 crc kubenswrapper[4894]: E1209 15:32:08.768546 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:12.76854004 +0000 UTC m=+27.087750799 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.771349 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.816547 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.840538 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.840577 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.840586 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.840600 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.840610 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:08Z","lastTransitionTime":"2025-12-09T15:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.850892 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.870017 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.870071 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:08 crc kubenswrapper[4894]: E1209 15:32:08.870222 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 15:32:08 crc kubenswrapper[4894]: E1209 15:32:08.870252 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 15:32:08 crc kubenswrapper[4894]: E1209 15:32:08.870262 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:08 crc kubenswrapper[4894]: E1209 15:32:08.870273 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 15:32:08 crc kubenswrapper[4894]: E1209 15:32:08.870294 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 15:32:08 crc kubenswrapper[4894]: E1209 15:32:08.870318 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:12.870301501 +0000 UTC m=+27.189512170 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:08 crc kubenswrapper[4894]: E1209 15:32:08.870320 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:08 crc kubenswrapper[4894]: E1209 15:32:08.870388 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:12.870369573 +0000 UTC m=+27.189580242 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.891873 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.930236 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.943194 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.943237 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.943248 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.943263 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.943275 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:08Z","lastTransitionTime":"2025-12-09T15:32:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:08 crc kubenswrapper[4894]: I1209 15:32:08.972899 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.013066 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:09Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.045592 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.045656 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.045668 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.045684 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.045693 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:09Z","lastTransitionTime":"2025-12-09T15:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.051412 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:09Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.105798 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.105873 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:09 crc kubenswrapper[4894]: E1209 15:32:09.105933 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:09 crc kubenswrapper[4894]: E1209 15:32:09.106008 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.107811 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:09Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.128723 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:09Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.148241 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.148273 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.148282 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.148295 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.148305 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:09Z","lastTransitionTime":"2025-12-09T15:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.251060 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.251104 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.251116 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.251132 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.251147 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:09Z","lastTransitionTime":"2025-12-09T15:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.305501 4894 generic.go:334] "Generic (PLEG): container finished" podID="05e6de1b-a262-4d89-ad3b-902a436d44cc" containerID="075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017" exitCode=0 Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.305784 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" event={"ID":"05e6de1b-a262-4d89-ad3b-902a436d44cc","Type":"ContainerDied","Data":"075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017"} Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.319775 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:09Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.330580 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:09Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.342650 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:09Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.353366 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:09Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.353758 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.353794 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.353806 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.353823 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.353834 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:09Z","lastTransitionTime":"2025-12-09T15:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.363085 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:09Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.374177 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:09Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.410758 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:09Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.449568 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:09Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.455736 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.455770 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.455778 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.455792 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.455801 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:09Z","lastTransitionTime":"2025-12-09T15:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.491536 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:09Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.540450 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:09Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.558472 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.558533 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.558556 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.558586 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.558610 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:09Z","lastTransitionTime":"2025-12-09T15:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.578987 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:09Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.610689 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:09Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.651490 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:09Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.661064 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.661103 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.661115 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.661130 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.661154 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:09Z","lastTransitionTime":"2025-12-09T15:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.690741 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:09Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.763094 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.763127 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.763138 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.763150 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.763160 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:09Z","lastTransitionTime":"2025-12-09T15:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.865599 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.865648 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.865665 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.865681 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.865692 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:09Z","lastTransitionTime":"2025-12-09T15:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.967704 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.967743 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.967754 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.967769 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:09 crc kubenswrapper[4894]: I1209 15:32:09.967782 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:09Z","lastTransitionTime":"2025-12-09T15:32:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.070225 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.070275 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.070290 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.070313 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.070329 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:10Z","lastTransitionTime":"2025-12-09T15:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.105985 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:10 crc kubenswrapper[4894]: E1209 15:32:10.106124 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.172576 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.172652 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.172665 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.172681 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.172691 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:10Z","lastTransitionTime":"2025-12-09T15:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.275538 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.275582 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.275595 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.275613 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.275626 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:10Z","lastTransitionTime":"2025-12-09T15:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.310517 4894 generic.go:334] "Generic (PLEG): container finished" podID="05e6de1b-a262-4d89-ad3b-902a436d44cc" containerID="e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792" exitCode=0 Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.310563 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" event={"ID":"05e6de1b-a262-4d89-ad3b-902a436d44cc","Type":"ContainerDied","Data":"e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792"} Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.323248 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:10Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.334392 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:10Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.348940 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:10Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.363350 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:10Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.379976 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.380032 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.380042 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.380059 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.380070 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:10Z","lastTransitionTime":"2025-12-09T15:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.381344 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:10Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.394192 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:10Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.408498 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:10Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.431712 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:10Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.449414 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:10Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.464439 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:10Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.478179 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:10Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.482207 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.482231 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.482239 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.482252 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.482260 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:10Z","lastTransitionTime":"2025-12-09T15:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.495477 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:10Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.507332 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:10Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.517368 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:10Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.584205 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.584237 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.584248 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.584263 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.584274 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:10Z","lastTransitionTime":"2025-12-09T15:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.687356 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.687742 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.687754 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.687772 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.687784 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:10Z","lastTransitionTime":"2025-12-09T15:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.791070 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.791109 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.791122 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.791139 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.791152 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:10Z","lastTransitionTime":"2025-12-09T15:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.893527 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.893579 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.893593 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.893612 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.893624 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:10Z","lastTransitionTime":"2025-12-09T15:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.996419 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.996459 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.996471 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.996491 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:10 crc kubenswrapper[4894]: I1209 15:32:10.996505 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:10Z","lastTransitionTime":"2025-12-09T15:32:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.098839 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.098892 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.098902 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.098918 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.098930 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:11Z","lastTransitionTime":"2025-12-09T15:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.106284 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:11 crc kubenswrapper[4894]: E1209 15:32:11.106372 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.106625 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:11 crc kubenswrapper[4894]: E1209 15:32:11.106860 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.221882 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.221959 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.221969 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.221984 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.221993 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:11Z","lastTransitionTime":"2025-12-09T15:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.317691 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerStarted","Data":"8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0"} Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.321687 4894 generic.go:334] "Generic (PLEG): container finished" podID="05e6de1b-a262-4d89-ad3b-902a436d44cc" containerID="355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317" exitCode=0 Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.321729 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" event={"ID":"05e6de1b-a262-4d89-ad3b-902a436d44cc","Type":"ContainerDied","Data":"355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317"} Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.323844 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.323903 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.323925 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.323952 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.323973 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:11Z","lastTransitionTime":"2025-12-09T15:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.372040 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:11Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.405038 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:11Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.426003 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.426029 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.426037 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.426050 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.426058 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:11Z","lastTransitionTime":"2025-12-09T15:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.428819 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:11Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.442318 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:11Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.453449 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:11Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.466900 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:11Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.478744 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:11Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.489843 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:11Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.498696 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:11Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.511682 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:11Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.527911 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.527935 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.527943 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.527956 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.527965 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:11Z","lastTransitionTime":"2025-12-09T15:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.530415 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:11Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.542921 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:11Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.555079 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:11Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.567196 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:11Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.630180 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.630248 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.630263 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.630280 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.630319 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:11Z","lastTransitionTime":"2025-12-09T15:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.732851 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.732889 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.732900 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.732916 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.732927 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:11Z","lastTransitionTime":"2025-12-09T15:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.836037 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.836085 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.836096 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.836113 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.836125 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:11Z","lastTransitionTime":"2025-12-09T15:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.938406 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.938444 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.938455 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.938470 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:11 crc kubenswrapper[4894]: I1209 15:32:11.938482 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:11Z","lastTransitionTime":"2025-12-09T15:32:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.040445 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.040486 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.040495 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.040509 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.040518 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:12Z","lastTransitionTime":"2025-12-09T15:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.106505 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:12 crc kubenswrapper[4894]: E1209 15:32:12.106704 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.143154 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.143195 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.143204 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.143218 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.143229 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:12Z","lastTransitionTime":"2025-12-09T15:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.245867 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.245902 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.245912 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.245929 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.245939 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:12Z","lastTransitionTime":"2025-12-09T15:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.329434 4894 generic.go:334] "Generic (PLEG): container finished" podID="05e6de1b-a262-4d89-ad3b-902a436d44cc" containerID="880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f" exitCode=0 Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.329484 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" event={"ID":"05e6de1b-a262-4d89-ad3b-902a436d44cc","Type":"ContainerDied","Data":"880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f"} Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.345902 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:12Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.348534 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.349081 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.349098 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.349117 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.349129 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:12Z","lastTransitionTime":"2025-12-09T15:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.362858 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:12Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.382990 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:12Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.402071 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:12Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.429263 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:12Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.447335 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:12Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.451386 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.451421 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.451431 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.451446 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.451457 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:12Z","lastTransitionTime":"2025-12-09T15:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.459091 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:12Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.471510 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:12Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.483019 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:12Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.498820 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:12Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.511294 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:12Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.524799 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:12Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.538746 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:12Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.552266 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:12Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.553818 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.553858 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.553870 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.553889 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.553902 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:12Z","lastTransitionTime":"2025-12-09T15:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.659742 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.659770 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.659780 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.659793 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.659803 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:12Z","lastTransitionTime":"2025-12-09T15:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.762174 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.762221 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.762236 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.762259 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.762274 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:12Z","lastTransitionTime":"2025-12-09T15:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.838668 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.838773 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.838797 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:12 crc kubenswrapper[4894]: E1209 15:32:12.838886 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 15:32:12 crc kubenswrapper[4894]: E1209 15:32:12.838907 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:32:20.838880729 +0000 UTC m=+35.158091388 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:32:12 crc kubenswrapper[4894]: E1209 15:32:12.838955 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:20.838934991 +0000 UTC m=+35.158145660 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 15:32:12 crc kubenswrapper[4894]: E1209 15:32:12.839060 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 15:32:12 crc kubenswrapper[4894]: E1209 15:32:12.839149 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:20.839130926 +0000 UTC m=+35.158341675 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.864219 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.864261 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.864274 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.864290 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.864301 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:12Z","lastTransitionTime":"2025-12-09T15:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.940470 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.940555 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:12 crc kubenswrapper[4894]: E1209 15:32:12.940661 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 15:32:12 crc kubenswrapper[4894]: E1209 15:32:12.940694 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 15:32:12 crc kubenswrapper[4894]: E1209 15:32:12.940706 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:12 crc kubenswrapper[4894]: E1209 15:32:12.940759 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:20.940743883 +0000 UTC m=+35.259954552 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:12 crc kubenswrapper[4894]: E1209 15:32:12.940810 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 15:32:12 crc kubenswrapper[4894]: E1209 15:32:12.940846 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 15:32:12 crc kubenswrapper[4894]: E1209 15:32:12.940858 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:12 crc kubenswrapper[4894]: E1209 15:32:12.940919 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:20.940902507 +0000 UTC m=+35.260113176 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.966782 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.966829 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.966841 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.966860 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:12 crc kubenswrapper[4894]: I1209 15:32:12.966874 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:12Z","lastTransitionTime":"2025-12-09T15:32:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.069669 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.069704 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.069715 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.069730 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.069742 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:13Z","lastTransitionTime":"2025-12-09T15:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.105369 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.105377 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:13 crc kubenswrapper[4894]: E1209 15:32:13.105527 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:13 crc kubenswrapper[4894]: E1209 15:32:13.105581 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.172380 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.172430 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.172444 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.172461 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.172474 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:13Z","lastTransitionTime":"2025-12-09T15:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.275754 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.275805 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.275819 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.275836 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.275847 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:13Z","lastTransitionTime":"2025-12-09T15:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.338910 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" event={"ID":"05e6de1b-a262-4d89-ad3b-902a436d44cc","Type":"ContainerStarted","Data":"6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8"} Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.344179 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerStarted","Data":"97a389993b82a942cdf2aa421ee41b672feb123ea942b2623bbd47a3477d4f06"} Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.344433 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.344457 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.359376 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.378419 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.378450 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.378459 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.378476 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.378490 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:13Z","lastTransitionTime":"2025-12-09T15:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.412474 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.412547 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.414247 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.424875 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.437366 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.451229 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.478410 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.481549 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.481592 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.481601 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.481616 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.481629 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:13Z","lastTransitionTime":"2025-12-09T15:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.490661 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.504241 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.516414 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.537141 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.553491 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.566428 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.580323 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.584364 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.584409 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.584424 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.584444 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.584457 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:13Z","lastTransitionTime":"2025-12-09T15:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.591027 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.602286 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.613751 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.624250 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.632977 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.644191 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.657694 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.669352 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.682945 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.687767 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.687801 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.687811 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.687829 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.687840 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:13Z","lastTransitionTime":"2025-12-09T15:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.696151 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.709507 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.727968 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97a389993b82a942cdf2aa421ee41b672feb123ea942b2623bbd47a3477d4f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.741125 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.751447 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.760848 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:13Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.791598 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.791681 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.791694 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.791714 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.791730 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:13Z","lastTransitionTime":"2025-12-09T15:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.894181 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.894255 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.894270 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.894291 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.894309 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:13Z","lastTransitionTime":"2025-12-09T15:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.997777 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.997866 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.997891 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.997923 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:13 crc kubenswrapper[4894]: I1209 15:32:13.997944 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:13Z","lastTransitionTime":"2025-12-09T15:32:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.100452 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.100489 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.100499 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.100513 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.100526 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:14Z","lastTransitionTime":"2025-12-09T15:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.105860 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:14 crc kubenswrapper[4894]: E1209 15:32:14.105969 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.202897 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.202941 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.202952 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.202970 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.202982 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:14Z","lastTransitionTime":"2025-12-09T15:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.305197 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.305236 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.305248 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.305272 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.305285 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:14Z","lastTransitionTime":"2025-12-09T15:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.346663 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.408384 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.408455 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.408472 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.408497 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.408514 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:14Z","lastTransitionTime":"2025-12-09T15:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.511097 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.511136 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.511155 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.511172 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.511207 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:14Z","lastTransitionTime":"2025-12-09T15:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.614515 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.614944 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.614959 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.614979 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.614991 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:14Z","lastTransitionTime":"2025-12-09T15:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.717679 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.717720 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.717729 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.717743 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.717752 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:14Z","lastTransitionTime":"2025-12-09T15:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.820317 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.820357 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.820366 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.820381 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.820393 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:14Z","lastTransitionTime":"2025-12-09T15:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.922273 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.922317 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.922332 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.922348 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:14 crc kubenswrapper[4894]: I1209 15:32:14.922358 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:14Z","lastTransitionTime":"2025-12-09T15:32:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.024627 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.024714 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.024726 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.024745 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.024758 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:15Z","lastTransitionTime":"2025-12-09T15:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.105775 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.105783 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:15 crc kubenswrapper[4894]: E1209 15:32:15.105919 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:15 crc kubenswrapper[4894]: E1209 15:32:15.105998 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.128501 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.128561 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.128576 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.128597 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.128610 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:15Z","lastTransitionTime":"2025-12-09T15:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.231188 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.231227 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.231238 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.231252 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.231265 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:15Z","lastTransitionTime":"2025-12-09T15:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.337135 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.337188 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.337209 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.337226 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.337238 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:15Z","lastTransitionTime":"2025-12-09T15:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.351129 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s5scr_97036c6f-dce3-444b-85e6-8476c16613c9/ovnkube-controller/0.log" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.354249 4894 generic.go:334] "Generic (PLEG): container finished" podID="97036c6f-dce3-444b-85e6-8476c16613c9" containerID="97a389993b82a942cdf2aa421ee41b672feb123ea942b2623bbd47a3477d4f06" exitCode=1 Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.354336 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerDied","Data":"97a389993b82a942cdf2aa421ee41b672feb123ea942b2623bbd47a3477d4f06"} Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.355444 4894 scope.go:117] "RemoveContainer" containerID="97a389993b82a942cdf2aa421ee41b672feb123ea942b2623bbd47a3477d4f06" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.373020 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.391044 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.405132 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.417200 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.431179 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.441695 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.441733 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.441745 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.441760 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.441772 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:15Z","lastTransitionTime":"2025-12-09T15:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.444669 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.457988 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.467304 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.478829 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.489517 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.498967 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.508964 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.520115 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.569501 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.569534 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.569543 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.569558 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.569569 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:15Z","lastTransitionTime":"2025-12-09T15:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.572390 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97a389993b82a942cdf2aa421ee41b672feb123ea942b2623bbd47a3477d4f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97a389993b82a942cdf2aa421ee41b672feb123ea942b2623bbd47a3477d4f06\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:15Z\\\",\\\"message\\\":\\\"etes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 15:32:14.740519 6169 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1209 15:32:14.741062 6169 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1209 15:32:14.741093 6169 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1209 15:32:14.741100 6169 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1209 15:32:14.741122 6169 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1209 15:32:14.741127 6169 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1209 15:32:14.741136 6169 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 15:32:14.741141 6169 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 15:32:14.741149 6169 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1209 15:32:14.741161 6169 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 15:32:14.741202 6169 factory.go:656] Stopping watch factory\\\\nI1209 15:32:14.741219 6169 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 15:32:14.741262 6169 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:15Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.671620 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.671744 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.671758 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.671784 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.671800 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:15Z","lastTransitionTime":"2025-12-09T15:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.775263 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.775320 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.775331 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.775353 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.775365 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:15Z","lastTransitionTime":"2025-12-09T15:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.878078 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.878135 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.878151 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.878172 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.878183 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:15Z","lastTransitionTime":"2025-12-09T15:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.980334 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.980384 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.980399 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.980413 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:15 crc kubenswrapper[4894]: I1209 15:32:15.980423 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:15Z","lastTransitionTime":"2025-12-09T15:32:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.082604 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.082662 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.082676 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.082697 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.082709 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:16Z","lastTransitionTime":"2025-12-09T15:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.105925 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:16 crc kubenswrapper[4894]: E1209 15:32:16.106070 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.130806 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://97a389993b82a942cdf2aa421ee41b672feb123ea942b2623bbd47a3477d4f06\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97a389993b82a942cdf2aa421ee41b672feb123ea942b2623bbd47a3477d4f06\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:15Z\\\",\\\"message\\\":\\\"etes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 15:32:14.740519 6169 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1209 15:32:14.741062 6169 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1209 15:32:14.741093 6169 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1209 15:32:14.741100 6169 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1209 15:32:14.741122 6169 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1209 15:32:14.741127 6169 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1209 15:32:14.741136 6169 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 15:32:14.741141 6169 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 15:32:14.741149 6169 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1209 15:32:14.741161 6169 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 15:32:14.741202 6169 factory.go:656] Stopping watch factory\\\\nI1209 15:32:14.741219 6169 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 15:32:14.741262 6169 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.142820 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.152473 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.161549 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.177796 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.185520 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.185572 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.185582 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.185597 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.185607 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:16Z","lastTransitionTime":"2025-12-09T15:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.197703 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.210815 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.225704 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.239705 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.261233 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.273657 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.286427 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.288240 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.288291 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.288304 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.288323 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.288335 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:16Z","lastTransitionTime":"2025-12-09T15:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.296153 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.306554 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.359597 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s5scr_97036c6f-dce3-444b-85e6-8476c16613c9/ovnkube-controller/0.log" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.362128 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerStarted","Data":"fbc99dc4587bee2241be7ff18f707d60bc3250e5c718d23a09884973cd22b24f"} Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.362247 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.375258 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.387318 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.390822 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.390860 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.390871 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.390885 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.390896 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:16Z","lastTransitionTime":"2025-12-09T15:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.397707 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.411689 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.429532 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc99dc4587bee2241be7ff18f707d60bc3250e5c718d23a09884973cd22b24f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97a389993b82a942cdf2aa421ee41b672feb123ea942b2623bbd47a3477d4f06\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:15Z\\\",\\\"message\\\":\\\"etes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 15:32:14.740519 6169 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1209 15:32:14.741062 6169 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1209 15:32:14.741093 6169 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1209 15:32:14.741100 6169 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1209 15:32:14.741122 6169 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1209 15:32:14.741127 6169 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1209 15:32:14.741136 6169 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 15:32:14.741141 6169 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 15:32:14.741149 6169 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1209 15:32:14.741161 6169 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 15:32:14.741202 6169 factory.go:656] Stopping watch factory\\\\nI1209 15:32:14.741219 6169 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 15:32:14.741262 6169 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.444111 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.454687 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.466790 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.476809 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.488826 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.493021 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.493073 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.493084 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.493103 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.493115 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:16Z","lastTransitionTime":"2025-12-09T15:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.500565 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.511751 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.520284 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.530414 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:16Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.596317 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.596356 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.596368 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.596389 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.596401 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:16Z","lastTransitionTime":"2025-12-09T15:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.699865 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.699898 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.699907 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.699921 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.699930 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:16Z","lastTransitionTime":"2025-12-09T15:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.803197 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.803278 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.803305 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.803332 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.803352 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:16Z","lastTransitionTime":"2025-12-09T15:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.906820 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.906892 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.906911 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.906937 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:16 crc kubenswrapper[4894]: I1209 15:32:16.906959 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:16Z","lastTransitionTime":"2025-12-09T15:32:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.010417 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.010491 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.010508 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.010533 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.010551 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:17Z","lastTransitionTime":"2025-12-09T15:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.106452 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.106519 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:17 crc kubenswrapper[4894]: E1209 15:32:17.106731 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:17 crc kubenswrapper[4894]: E1209 15:32:17.106835 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.112946 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.113014 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.113026 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.113067 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.113081 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:17Z","lastTransitionTime":"2025-12-09T15:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.215438 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.215485 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.215500 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.215519 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.215532 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:17Z","lastTransitionTime":"2025-12-09T15:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.318276 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.318317 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.318325 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.318340 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.318349 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:17Z","lastTransitionTime":"2025-12-09T15:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.365158 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.421454 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.421507 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.421527 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.421551 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.421568 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:17Z","lastTransitionTime":"2025-12-09T15:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.433843 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62"] Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.434248 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.437009 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.437077 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.448396 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.461184 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.472538 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.485170 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1473dc3d-be11-41e4-88d4-995a80be1e53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gsk62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.493156 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.493207 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.493222 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.493242 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.493274 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1473dc3d-be11-41e4-88d4-995a80be1e53-env-overrides\") pod \"ovnkube-control-plane-749d76644c-gsk62\" (UID: \"1473dc3d-be11-41e4-88d4-995a80be1e53\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.493355 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1473dc3d-be11-41e4-88d4-995a80be1e53-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-gsk62\" (UID: \"1473dc3d-be11-41e4-88d4-995a80be1e53\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.493386 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r9nm5\" (UniqueName: \"kubernetes.io/projected/1473dc3d-be11-41e4-88d4-995a80be1e53-kube-api-access-r9nm5\") pod \"ovnkube-control-plane-749d76644c-gsk62\" (UID: \"1473dc3d-be11-41e4-88d4-995a80be1e53\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.493406 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1473dc3d-be11-41e4-88d4-995a80be1e53-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-gsk62\" (UID: \"1473dc3d-be11-41e4-88d4-995a80be1e53\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.493299 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:17Z","lastTransitionTime":"2025-12-09T15:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.499467 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:17 crc kubenswrapper[4894]: E1209 15:32:17.505919 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.509744 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.509769 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.509777 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.509791 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.509799 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:17Z","lastTransitionTime":"2025-12-09T15:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.515804 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:17 crc kubenswrapper[4894]: E1209 15:32:17.525354 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.528865 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.528981 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.529047 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.529123 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.529191 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:17Z","lastTransitionTime":"2025-12-09T15:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.530180 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.541060 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:17 crc kubenswrapper[4894]: E1209 15:32:17.541197 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.544930 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.544967 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.544976 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.544991 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.545000 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:17Z","lastTransitionTime":"2025-12-09T15:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.562268 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.593715 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.594377 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r9nm5\" (UniqueName: \"kubernetes.io/projected/1473dc3d-be11-41e4-88d4-995a80be1e53-kube-api-access-r9nm5\") pod \"ovnkube-control-plane-749d76644c-gsk62\" (UID: \"1473dc3d-be11-41e4-88d4-995a80be1e53\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.594412 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1473dc3d-be11-41e4-88d4-995a80be1e53-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-gsk62\" (UID: \"1473dc3d-be11-41e4-88d4-995a80be1e53\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.594553 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1473dc3d-be11-41e4-88d4-995a80be1e53-env-overrides\") pod \"ovnkube-control-plane-749d76644c-gsk62\" (UID: \"1473dc3d-be11-41e4-88d4-995a80be1e53\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.594624 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1473dc3d-be11-41e4-88d4-995a80be1e53-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-gsk62\" (UID: \"1473dc3d-be11-41e4-88d4-995a80be1e53\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.595535 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/1473dc3d-be11-41e4-88d4-995a80be1e53-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-gsk62\" (UID: \"1473dc3d-be11-41e4-88d4-995a80be1e53\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.595693 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/1473dc3d-be11-41e4-88d4-995a80be1e53-env-overrides\") pod \"ovnkube-control-plane-749d76644c-gsk62\" (UID: \"1473dc3d-be11-41e4-88d4-995a80be1e53\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" Dec 09 15:32:17 crc kubenswrapper[4894]: E1209 15:32:17.604927 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.606987 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/1473dc3d-be11-41e4-88d4-995a80be1e53-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-gsk62\" (UID: \"1473dc3d-be11-41e4-88d4-995a80be1e53\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.616089 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.616149 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.616164 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.616202 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.616216 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:17Z","lastTransitionTime":"2025-12-09T15:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.625417 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r9nm5\" (UniqueName: \"kubernetes.io/projected/1473dc3d-be11-41e4-88d4-995a80be1e53-kube-api-access-r9nm5\") pod \"ovnkube-control-plane-749d76644c-gsk62\" (UID: \"1473dc3d-be11-41e4-88d4-995a80be1e53\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.627140 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:17 crc kubenswrapper[4894]: E1209 15:32:17.635248 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:17 crc kubenswrapper[4894]: E1209 15:32:17.635404 4894 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.637256 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.637281 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.637310 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.637325 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.637333 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:17Z","lastTransitionTime":"2025-12-09T15:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.644361 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.654207 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.668586 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.685715 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc99dc4587bee2241be7ff18f707d60bc3250e5c718d23a09884973cd22b24f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97a389993b82a942cdf2aa421ee41b672feb123ea942b2623bbd47a3477d4f06\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:15Z\\\",\\\"message\\\":\\\"etes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 15:32:14.740519 6169 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1209 15:32:14.741062 6169 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1209 15:32:14.741093 6169 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1209 15:32:14.741100 6169 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1209 15:32:14.741122 6169 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1209 15:32:14.741127 6169 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1209 15:32:14.741136 6169 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 15:32:14.741141 6169 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 15:32:14.741149 6169 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1209 15:32:14.741161 6169 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 15:32:14.741202 6169 factory.go:656] Stopping watch factory\\\\nI1209 15:32:14.741219 6169 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 15:32:14.741262 6169 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:17Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.740216 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.740577 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.740590 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.740607 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.740657 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:17Z","lastTransitionTime":"2025-12-09T15:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.746473 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.843014 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.843044 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.843054 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.843067 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.843076 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:17Z","lastTransitionTime":"2025-12-09T15:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.945286 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.945320 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.945332 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.945348 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:17 crc kubenswrapper[4894]: I1209 15:32:17.945359 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:17Z","lastTransitionTime":"2025-12-09T15:32:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.048297 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.048336 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.048348 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.048364 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.048376 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:18Z","lastTransitionTime":"2025-12-09T15:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.106002 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:18 crc kubenswrapper[4894]: E1209 15:32:18.106170 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.150843 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.150875 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.150883 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.150895 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.150904 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:18Z","lastTransitionTime":"2025-12-09T15:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.253648 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.253710 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.253721 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.253736 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.253746 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:18Z","lastTransitionTime":"2025-12-09T15:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.358163 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.358218 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.358235 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.358259 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.358276 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:18Z","lastTransitionTime":"2025-12-09T15:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.371808 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s5scr_97036c6f-dce3-444b-85e6-8476c16613c9/ovnkube-controller/1.log" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.373955 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s5scr_97036c6f-dce3-444b-85e6-8476c16613c9/ovnkube-controller/0.log" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.377895 4894 generic.go:334] "Generic (PLEG): container finished" podID="97036c6f-dce3-444b-85e6-8476c16613c9" containerID="fbc99dc4587bee2241be7ff18f707d60bc3250e5c718d23a09884973cd22b24f" exitCode=1 Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.377959 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerDied","Data":"fbc99dc4587bee2241be7ff18f707d60bc3250e5c718d23a09884973cd22b24f"} Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.378008 4894 scope.go:117] "RemoveContainer" containerID="97a389993b82a942cdf2aa421ee41b672feb123ea942b2623bbd47a3477d4f06" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.379667 4894 scope.go:117] "RemoveContainer" containerID="fbc99dc4587bee2241be7ff18f707d60bc3250e5c718d23a09884973cd22b24f" Dec 09 15:32:18 crc kubenswrapper[4894]: E1209 15:32:18.380041 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-s5scr_openshift-ovn-kubernetes(97036c6f-dce3-444b-85e6-8476c16613c9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.380807 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" event={"ID":"1473dc3d-be11-41e4-88d4-995a80be1e53","Type":"ContainerStarted","Data":"0a183cf558c984ca981afe6cfbeda1cb173a511751d8edb975cf87e677e55c68"} Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.380857 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" event={"ID":"1473dc3d-be11-41e4-88d4-995a80be1e53","Type":"ContainerStarted","Data":"23725ac5b4972866e4473ac7e2a53c40bb6bc662bec277dd1f4f0d4e4feb2262"} Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.395923 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:18Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.410013 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:18Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.460537 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.460581 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.460598 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.460619 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.460663 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:18Z","lastTransitionTime":"2025-12-09T15:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.473548 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:18Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.485967 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:18Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.497039 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:18Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.509750 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:18Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.522005 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:18Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.533311 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:18Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.550150 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:18Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.563183 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.563223 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.563235 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.563253 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.563265 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:18Z","lastTransitionTime":"2025-12-09T15:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.572179 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc99dc4587bee2241be7ff18f707d60bc3250e5c718d23a09884973cd22b24f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97a389993b82a942cdf2aa421ee41b672feb123ea942b2623bbd47a3477d4f06\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:15Z\\\",\\\"message\\\":\\\"etes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 15:32:14.740519 6169 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1209 15:32:14.741062 6169 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1209 15:32:14.741093 6169 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1209 15:32:14.741100 6169 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1209 15:32:14.741122 6169 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1209 15:32:14.741127 6169 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1209 15:32:14.741136 6169 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 15:32:14.741141 6169 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 15:32:14.741149 6169 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1209 15:32:14.741161 6169 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 15:32:14.741202 6169 factory.go:656] Stopping watch factory\\\\nI1209 15:32:14.741219 6169 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 15:32:14.741262 6169 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc99dc4587bee2241be7ff18f707d60bc3250e5c718d23a09884973cd22b24f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"ift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1209 15:32:16.221102 6296 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 15:32:16.221118 6296 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1209 15:32:16.221130 6296 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1209 15:32:16.221193 6296 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:18Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.585922 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:18Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.598014 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:18Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.612778 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:18Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.624675 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1473dc3d-be11-41e4-88d4-995a80be1e53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gsk62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:18Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.638260 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:18Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.666236 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.666303 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.666313 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.666327 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.666337 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:18Z","lastTransitionTime":"2025-12-09T15:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.768441 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.768485 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.768500 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.768515 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.768525 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:18Z","lastTransitionTime":"2025-12-09T15:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.870202 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.870291 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.870303 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.870320 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.870331 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:18Z","lastTransitionTime":"2025-12-09T15:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.929240 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-qxfr8"] Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.930021 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:18 crc kubenswrapper[4894]: E1209 15:32:18.930129 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.941600 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:18Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.953024 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:18Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.962349 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:18Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.973169 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.973225 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.973234 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.973247 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.973256 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:18Z","lastTransitionTime":"2025-12-09T15:32:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.973948 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:18Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.984039 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qxfr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"948b28c5-e35c-4e6c-b31c-c0d5b205389b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qxfr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:18Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:18 crc kubenswrapper[4894]: I1209 15:32:18.996540 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:18Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.007563 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.008813 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bggtq\" (UniqueName: \"kubernetes.io/projected/948b28c5-e35c-4e6c-b31c-c0d5b205389b-kube-api-access-bggtq\") pod \"network-metrics-daemon-qxfr8\" (UID: \"948b28c5-e35c-4e6c-b31c-c0d5b205389b\") " pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.008878 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs\") pod \"network-metrics-daemon-qxfr8\" (UID: \"948b28c5-e35c-4e6c-b31c-c0d5b205389b\") " pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.018485 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.028220 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.044908 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.063821 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc99dc4587bee2241be7ff18f707d60bc3250e5c718d23a09884973cd22b24f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97a389993b82a942cdf2aa421ee41b672feb123ea942b2623bbd47a3477d4f06\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:15Z\\\",\\\"message\\\":\\\"etes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 15:32:14.740519 6169 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1209 15:32:14.741062 6169 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1209 15:32:14.741093 6169 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1209 15:32:14.741100 6169 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1209 15:32:14.741122 6169 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1209 15:32:14.741127 6169 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1209 15:32:14.741136 6169 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 15:32:14.741141 6169 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 15:32:14.741149 6169 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1209 15:32:14.741161 6169 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 15:32:14.741202 6169 factory.go:656] Stopping watch factory\\\\nI1209 15:32:14.741219 6169 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 15:32:14.741262 6169 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc99dc4587bee2241be7ff18f707d60bc3250e5c718d23a09884973cd22b24f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"ift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1209 15:32:16.221102 6296 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 15:32:16.221118 6296 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1209 15:32:16.221130 6296 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1209 15:32:16.221193 6296 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.075252 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.075285 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.075294 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.075306 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.075315 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:19Z","lastTransitionTime":"2025-12-09T15:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.076956 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.092919 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.105882 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.106048 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:19 crc kubenswrapper[4894]: E1209 15:32:19.106143 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:19 crc kubenswrapper[4894]: E1209 15:32:19.106301 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.109628 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bggtq\" (UniqueName: \"kubernetes.io/projected/948b28c5-e35c-4e6c-b31c-c0d5b205389b-kube-api-access-bggtq\") pod \"network-metrics-daemon-qxfr8\" (UID: \"948b28c5-e35c-4e6c-b31c-c0d5b205389b\") " pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.109700 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs\") pod \"network-metrics-daemon-qxfr8\" (UID: \"948b28c5-e35c-4e6c-b31c-c0d5b205389b\") " pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:19 crc kubenswrapper[4894]: E1209 15:32:19.109808 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 15:32:19 crc kubenswrapper[4894]: E1209 15:32:19.109863 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs podName:948b28c5-e35c-4e6c-b31c-c0d5b205389b nodeName:}" failed. No retries permitted until 2025-12-09 15:32:19.609846978 +0000 UTC m=+33.929057647 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs") pod "network-metrics-daemon-qxfr8" (UID: "948b28c5-e35c-4e6c-b31c-c0d5b205389b") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.110144 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.128660 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1473dc3d-be11-41e4-88d4-995a80be1e53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gsk62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.133212 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bggtq\" (UniqueName: \"kubernetes.io/projected/948b28c5-e35c-4e6c-b31c-c0d5b205389b-kube-api-access-bggtq\") pod \"network-metrics-daemon-qxfr8\" (UID: \"948b28c5-e35c-4e6c-b31c-c0d5b205389b\") " pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.150566 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.177481 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.177538 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.177550 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.177569 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.177582 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:19Z","lastTransitionTime":"2025-12-09T15:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.279694 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.279722 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.279730 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.279744 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.279753 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:19Z","lastTransitionTime":"2025-12-09T15:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.382791 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.382828 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.382841 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.382858 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.382870 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:19Z","lastTransitionTime":"2025-12-09T15:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.387051 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" event={"ID":"1473dc3d-be11-41e4-88d4-995a80be1e53","Type":"ContainerStarted","Data":"100782eeaa80dd7da85d32c80957a78582580bdf2bc44281f52b87865cffbe35"} Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.389299 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s5scr_97036c6f-dce3-444b-85e6-8476c16613c9/ovnkube-controller/1.log" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.405513 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1473dc3d-be11-41e4-88d4-995a80be1e53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a183cf558c984ca981afe6cfbeda1cb173a511751d8edb975cf87e677e55c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://100782eeaa80dd7da85d32c80957a78582580bdf2bc44281f52b87865cffbe35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gsk62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.429284 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.449860 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.462850 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.472918 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.485719 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.485770 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.485784 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.485805 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.485817 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:19Z","lastTransitionTime":"2025-12-09T15:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.486684 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.496610 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qxfr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"948b28c5-e35c-4e6c-b31c-c0d5b205389b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qxfr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.512184 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.526947 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.541372 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.551998 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.565899 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.585232 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc99dc4587bee2241be7ff18f707d60bc3250e5c718d23a09884973cd22b24f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97a389993b82a942cdf2aa421ee41b672feb123ea942b2623bbd47a3477d4f06\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:15Z\\\",\\\"message\\\":\\\"etes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 15:32:14.740519 6169 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1209 15:32:14.741062 6169 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1209 15:32:14.741093 6169 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1209 15:32:14.741100 6169 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1209 15:32:14.741122 6169 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1209 15:32:14.741127 6169 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1209 15:32:14.741136 6169 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 15:32:14.741141 6169 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 15:32:14.741149 6169 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1209 15:32:14.741161 6169 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 15:32:14.741202 6169 factory.go:656] Stopping watch factory\\\\nI1209 15:32:14.741219 6169 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 15:32:14.741262 6169 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc99dc4587bee2241be7ff18f707d60bc3250e5c718d23a09884973cd22b24f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"ift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1209 15:32:16.221102 6296 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 15:32:16.221118 6296 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1209 15:32:16.221130 6296 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1209 15:32:16.221193 6296 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.588550 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.588586 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.588597 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.588612 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.588620 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:19Z","lastTransitionTime":"2025-12-09T15:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.596024 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.607416 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.614163 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs\") pod \"network-metrics-daemon-qxfr8\" (UID: \"948b28c5-e35c-4e6c-b31c-c0d5b205389b\") " pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:19 crc kubenswrapper[4894]: E1209 15:32:19.614301 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 15:32:19 crc kubenswrapper[4894]: E1209 15:32:19.614345 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs podName:948b28c5-e35c-4e6c-b31c-c0d5b205389b nodeName:}" failed. No retries permitted until 2025-12-09 15:32:20.614332339 +0000 UTC m=+34.933543008 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs") pod "network-metrics-daemon-qxfr8" (UID: "948b28c5-e35c-4e6c-b31c-c0d5b205389b") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.617434 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:19Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.691124 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.691189 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.691197 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.691211 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.691222 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:19Z","lastTransitionTime":"2025-12-09T15:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.792966 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.793003 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.793011 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.793027 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.793036 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:19Z","lastTransitionTime":"2025-12-09T15:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.895416 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.895461 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.895471 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.895483 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.895492 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:19Z","lastTransitionTime":"2025-12-09T15:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.997401 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.998129 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.998145 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.998161 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:19 crc kubenswrapper[4894]: I1209 15:32:19.998173 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:19Z","lastTransitionTime":"2025-12-09T15:32:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.100420 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.100476 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.100486 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.100506 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.100519 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:20Z","lastTransitionTime":"2025-12-09T15:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.105789 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:20 crc kubenswrapper[4894]: E1209 15:32:20.105949 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.203589 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.203664 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.203677 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.203692 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.203703 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:20Z","lastTransitionTime":"2025-12-09T15:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.305681 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.305723 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.305733 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.305748 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.305759 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:20Z","lastTransitionTime":"2025-12-09T15:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.408856 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.408885 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.408894 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.408906 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.408915 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:20Z","lastTransitionTime":"2025-12-09T15:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.511629 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.511696 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.511708 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.511724 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.511736 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:20Z","lastTransitionTime":"2025-12-09T15:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.614553 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.614714 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.614731 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.614758 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.614775 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:20Z","lastTransitionTime":"2025-12-09T15:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.624397 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs\") pod \"network-metrics-daemon-qxfr8\" (UID: \"948b28c5-e35c-4e6c-b31c-c0d5b205389b\") " pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:20 crc kubenswrapper[4894]: E1209 15:32:20.624562 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 15:32:20 crc kubenswrapper[4894]: E1209 15:32:20.624657 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs podName:948b28c5-e35c-4e6c-b31c-c0d5b205389b nodeName:}" failed. No retries permitted until 2025-12-09 15:32:22.624618227 +0000 UTC m=+36.943828896 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs") pod "network-metrics-daemon-qxfr8" (UID: "948b28c5-e35c-4e6c-b31c-c0d5b205389b") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.717766 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.717818 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.717832 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.717851 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.717864 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:20Z","lastTransitionTime":"2025-12-09T15:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.820714 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.820760 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.820772 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.820854 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.820870 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:20Z","lastTransitionTime":"2025-12-09T15:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.923557 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.923614 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.923625 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.923658 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.923670 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:20Z","lastTransitionTime":"2025-12-09T15:32:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.928299 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:32:20 crc kubenswrapper[4894]: E1209 15:32:20.928902 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:32:36.928861963 +0000 UTC m=+51.248072642 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.929098 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:20 crc kubenswrapper[4894]: I1209 15:32:20.929153 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:20 crc kubenswrapper[4894]: E1209 15:32:20.929187 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 15:32:20 crc kubenswrapper[4894]: E1209 15:32:20.929238 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:36.929226652 +0000 UTC m=+51.248437331 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 15:32:20 crc kubenswrapper[4894]: E1209 15:32:20.929243 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 15:32:20 crc kubenswrapper[4894]: E1209 15:32:20.929276 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:36.929263433 +0000 UTC m=+51.248474102 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.026008 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.026076 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.026098 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.026124 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.026147 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:21Z","lastTransitionTime":"2025-12-09T15:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.030442 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.030487 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:21 crc kubenswrapper[4894]: E1209 15:32:21.030630 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 15:32:21 crc kubenswrapper[4894]: E1209 15:32:21.030669 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 15:32:21 crc kubenswrapper[4894]: E1209 15:32:21.030682 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:21 crc kubenswrapper[4894]: E1209 15:32:21.030694 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 15:32:21 crc kubenswrapper[4894]: E1209 15:32:21.030732 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 15:32:21 crc kubenswrapper[4894]: E1209 15:32:21.030756 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:21 crc kubenswrapper[4894]: E1209 15:32:21.030734 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:37.030719947 +0000 UTC m=+51.349930626 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:21 crc kubenswrapper[4894]: E1209 15:32:21.030863 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 15:32:37.03083829 +0000 UTC m=+51.350048999 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.105557 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.105613 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:21 crc kubenswrapper[4894]: E1209 15:32:21.105713 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.105712 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:21 crc kubenswrapper[4894]: E1209 15:32:21.105851 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:21 crc kubenswrapper[4894]: E1209 15:32:21.105992 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.129326 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.129370 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.129382 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.129400 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.129411 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:21Z","lastTransitionTime":"2025-12-09T15:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.231540 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.231605 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.231617 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.231635 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.231663 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:21Z","lastTransitionTime":"2025-12-09T15:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.334291 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.334358 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.334371 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.334384 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.334393 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:21Z","lastTransitionTime":"2025-12-09T15:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.436457 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.436503 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.436512 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.436526 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.436536 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:21Z","lastTransitionTime":"2025-12-09T15:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.538964 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.539026 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.539038 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.539053 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.539064 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:21Z","lastTransitionTime":"2025-12-09T15:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.641309 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.641363 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.641372 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.641384 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.641393 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:21Z","lastTransitionTime":"2025-12-09T15:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.743423 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.743487 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.743499 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.743514 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.743523 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:21Z","lastTransitionTime":"2025-12-09T15:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.846533 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.846589 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.846604 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.846696 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.846716 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:21Z","lastTransitionTime":"2025-12-09T15:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.949341 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.949385 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.949394 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.949408 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:21 crc kubenswrapper[4894]: I1209 15:32:21.949417 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:21Z","lastTransitionTime":"2025-12-09T15:32:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.052326 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.052399 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.052413 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.052433 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.052453 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:22Z","lastTransitionTime":"2025-12-09T15:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.105982 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:22 crc kubenswrapper[4894]: E1209 15:32:22.106199 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.155236 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.155278 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.155288 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.155303 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.155314 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:22Z","lastTransitionTime":"2025-12-09T15:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.258294 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.258384 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.258407 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.258439 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.258457 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:22Z","lastTransitionTime":"2025-12-09T15:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.360401 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.360495 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.360512 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.360554 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.360566 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:22Z","lastTransitionTime":"2025-12-09T15:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.463403 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.463468 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.463486 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.463519 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.463556 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:22Z","lastTransitionTime":"2025-12-09T15:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.566050 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.566106 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.566118 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.566137 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.566150 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:22Z","lastTransitionTime":"2025-12-09T15:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.646935 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs\") pod \"network-metrics-daemon-qxfr8\" (UID: \"948b28c5-e35c-4e6c-b31c-c0d5b205389b\") " pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:22 crc kubenswrapper[4894]: E1209 15:32:22.647127 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 15:32:22 crc kubenswrapper[4894]: E1209 15:32:22.647201 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs podName:948b28c5-e35c-4e6c-b31c-c0d5b205389b nodeName:}" failed. No retries permitted until 2025-12-09 15:32:26.647182394 +0000 UTC m=+40.966393063 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs") pod "network-metrics-daemon-qxfr8" (UID: "948b28c5-e35c-4e6c-b31c-c0d5b205389b") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.668879 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.668916 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.668932 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.668948 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.668958 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:22Z","lastTransitionTime":"2025-12-09T15:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.771440 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.771483 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.771497 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.771513 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.771524 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:22Z","lastTransitionTime":"2025-12-09T15:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.873770 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.873818 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.873828 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.873842 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.873851 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:22Z","lastTransitionTime":"2025-12-09T15:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.976723 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.976779 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.976790 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.976805 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:22 crc kubenswrapper[4894]: I1209 15:32:22.976813 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:22Z","lastTransitionTime":"2025-12-09T15:32:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.118368 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.118480 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.118538 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:23 crc kubenswrapper[4894]: E1209 15:32:23.118584 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:23 crc kubenswrapper[4894]: E1209 15:32:23.118753 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:32:23 crc kubenswrapper[4894]: E1209 15:32:23.118865 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.119676 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.120616 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.120631 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.120668 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.120679 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:23Z","lastTransitionTime":"2025-12-09T15:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.223497 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.223539 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.223551 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.223566 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.223577 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:23Z","lastTransitionTime":"2025-12-09T15:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.337705 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.337767 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.337994 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.338028 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.338044 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:23Z","lastTransitionTime":"2025-12-09T15:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.441765 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.441829 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.441847 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.441871 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.441889 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:23Z","lastTransitionTime":"2025-12-09T15:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.545038 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.545079 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.545089 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.545105 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.545116 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:23Z","lastTransitionTime":"2025-12-09T15:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.647424 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.647454 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.647464 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.647480 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.647491 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:23Z","lastTransitionTime":"2025-12-09T15:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.749937 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.749968 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.749978 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.749997 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.750006 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:23Z","lastTransitionTime":"2025-12-09T15:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.853451 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.853510 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.853528 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.853552 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.853567 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:23Z","lastTransitionTime":"2025-12-09T15:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.956143 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.956198 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.956209 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.956224 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:23 crc kubenswrapper[4894]: I1209 15:32:23.956234 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:23Z","lastTransitionTime":"2025-12-09T15:32:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.058426 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.058465 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.058473 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.058487 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.058497 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:24Z","lastTransitionTime":"2025-12-09T15:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.105687 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:24 crc kubenswrapper[4894]: E1209 15:32:24.105873 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.160683 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.160729 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.160740 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.160754 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.160764 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:24Z","lastTransitionTime":"2025-12-09T15:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.263592 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.263657 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.263671 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.263684 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.263694 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:24Z","lastTransitionTime":"2025-12-09T15:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.366264 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.366316 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.366325 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.366342 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.366352 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:24Z","lastTransitionTime":"2025-12-09T15:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.468930 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.468986 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.469021 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.469066 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.469087 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:24Z","lastTransitionTime":"2025-12-09T15:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.571209 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.571251 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.571263 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.571279 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.571290 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:24Z","lastTransitionTime":"2025-12-09T15:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.673565 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.673920 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.673933 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.673954 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.673966 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:24Z","lastTransitionTime":"2025-12-09T15:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.776231 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.776269 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.776277 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.776292 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.776301 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:24Z","lastTransitionTime":"2025-12-09T15:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.800874 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.817289 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:24Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.832993 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:24Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.848440 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:24Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.861376 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1473dc3d-be11-41e4-88d4-995a80be1e53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a183cf558c984ca981afe6cfbeda1cb173a511751d8edb975cf87e677e55c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://100782eeaa80dd7da85d32c80957a78582580bdf2bc44281f52b87865cffbe35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gsk62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:24Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.873128 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:24Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.879086 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.879146 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.879166 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.879184 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.879194 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:24Z","lastTransitionTime":"2025-12-09T15:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.886892 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:24Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.900766 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:24Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.914805 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:24Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.925437 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:24Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.938635 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:24Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.950363 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qxfr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"948b28c5-e35c-4e6c-b31c-c0d5b205389b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qxfr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:24Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.962578 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:24Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.974623 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:24Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.982827 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.982855 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.982864 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.982878 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.982890 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:24Z","lastTransitionTime":"2025-12-09T15:32:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:24 crc kubenswrapper[4894]: I1209 15:32:24.986180 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:24Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.002878 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:25Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.021688 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc99dc4587bee2241be7ff18f707d60bc3250e5c718d23a09884973cd22b24f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97a389993b82a942cdf2aa421ee41b672feb123ea942b2623bbd47a3477d4f06\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:15Z\\\",\\\"message\\\":\\\"etes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 15:32:14.740519 6169 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1209 15:32:14.741062 6169 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1209 15:32:14.741093 6169 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1209 15:32:14.741100 6169 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1209 15:32:14.741122 6169 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1209 15:32:14.741127 6169 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1209 15:32:14.741136 6169 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 15:32:14.741141 6169 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 15:32:14.741149 6169 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1209 15:32:14.741161 6169 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 15:32:14.741202 6169 factory.go:656] Stopping watch factory\\\\nI1209 15:32:14.741219 6169 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 15:32:14.741262 6169 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc99dc4587bee2241be7ff18f707d60bc3250e5c718d23a09884973cd22b24f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"ift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1209 15:32:16.221102 6296 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 15:32:16.221118 6296 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1209 15:32:16.221130 6296 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1209 15:32:16.221193 6296 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:25Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.085409 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.085456 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.085466 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.085519 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.085531 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:25Z","lastTransitionTime":"2025-12-09T15:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.105812 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.105836 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:25 crc kubenswrapper[4894]: E1209 15:32:25.105903 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:25 crc kubenswrapper[4894]: E1209 15:32:25.105962 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.105841 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:25 crc kubenswrapper[4894]: E1209 15:32:25.106040 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.188445 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.188486 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.188495 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.188509 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.188518 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:25Z","lastTransitionTime":"2025-12-09T15:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.290580 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.290616 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.290625 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.290652 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.290662 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:25Z","lastTransitionTime":"2025-12-09T15:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.393197 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.393241 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.393257 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.393273 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.393284 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:25Z","lastTransitionTime":"2025-12-09T15:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.495618 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.495717 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.495732 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.495750 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.495761 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:25Z","lastTransitionTime":"2025-12-09T15:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.598661 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.598714 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.598727 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.598744 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.598755 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:25Z","lastTransitionTime":"2025-12-09T15:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.701026 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.701098 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.701111 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.701132 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.701144 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:25Z","lastTransitionTime":"2025-12-09T15:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.803119 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.803159 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.803168 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.803184 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.803196 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:25Z","lastTransitionTime":"2025-12-09T15:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.905316 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.905352 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.905362 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.905377 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:25 crc kubenswrapper[4894]: I1209 15:32:25.905387 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:25Z","lastTransitionTime":"2025-12-09T15:32:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.007980 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.008031 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.008043 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.008060 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.008071 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:26Z","lastTransitionTime":"2025-12-09T15:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.105913 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:26 crc kubenswrapper[4894]: E1209 15:32:26.106067 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.110208 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.110256 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.110267 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.110287 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.110304 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:26Z","lastTransitionTime":"2025-12-09T15:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.117585 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.127903 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qxfr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"948b28c5-e35c-4e6c-b31c-c0d5b205389b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qxfr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.143237 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.156148 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.170585 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.181746 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.192797 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.212127 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc99dc4587bee2241be7ff18f707d60bc3250e5c718d23a09884973cd22b24f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://97a389993b82a942cdf2aa421ee41b672feb123ea942b2623bbd47a3477d4f06\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:15Z\\\",\\\"message\\\":\\\"etes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1209 15:32:14.740519 6169 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1209 15:32:14.741062 6169 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI1209 15:32:14.741093 6169 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1209 15:32:14.741100 6169 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1209 15:32:14.741122 6169 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI1209 15:32:14.741127 6169 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1209 15:32:14.741136 6169 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1209 15:32:14.741141 6169 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1209 15:32:14.741149 6169 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1209 15:32:14.741161 6169 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 15:32:14.741202 6169 factory.go:656] Stopping watch factory\\\\nI1209 15:32:14.741219 6169 handler.go:208] Removed *v1.Node event handler 7\\\\nI1209 15:32:14.741262 6169 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc99dc4587bee2241be7ff18f707d60bc3250e5c718d23a09884973cd22b24f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"ift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1209 15:32:16.221102 6296 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 15:32:16.221118 6296 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1209 15:32:16.221130 6296 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1209 15:32:16.221193 6296 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.213062 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.213101 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.213114 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.213133 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.213145 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:26Z","lastTransitionTime":"2025-12-09T15:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.225240 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.239073 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.252824 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.269351 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.286907 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.300939 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.315402 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.315451 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.315463 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.315482 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.315496 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:26Z","lastTransitionTime":"2025-12-09T15:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.315407 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.327414 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1473dc3d-be11-41e4-88d4-995a80be1e53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a183cf558c984ca981afe6cfbeda1cb173a511751d8edb975cf87e677e55c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://100782eeaa80dd7da85d32c80957a78582580bdf2bc44281f52b87865cffbe35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gsk62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:26Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.418563 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.418610 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.418620 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.418684 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.418698 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:26Z","lastTransitionTime":"2025-12-09T15:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.520986 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.521023 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.521034 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.521049 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.521060 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:26Z","lastTransitionTime":"2025-12-09T15:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.623358 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.623394 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.623402 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.623418 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.623434 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:26Z","lastTransitionTime":"2025-12-09T15:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.689344 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs\") pod \"network-metrics-daemon-qxfr8\" (UID: \"948b28c5-e35c-4e6c-b31c-c0d5b205389b\") " pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:26 crc kubenswrapper[4894]: E1209 15:32:26.689506 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 15:32:26 crc kubenswrapper[4894]: E1209 15:32:26.689895 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs podName:948b28c5-e35c-4e6c-b31c-c0d5b205389b nodeName:}" failed. No retries permitted until 2025-12-09 15:32:34.689872577 +0000 UTC m=+49.009083246 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs") pod "network-metrics-daemon-qxfr8" (UID: "948b28c5-e35c-4e6c-b31c-c0d5b205389b") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.725443 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.725694 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.725795 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.725872 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.725954 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:26Z","lastTransitionTime":"2025-12-09T15:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.828500 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.828547 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.828558 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.828576 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.828588 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:26Z","lastTransitionTime":"2025-12-09T15:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.930954 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.931002 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.931017 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.931043 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:26 crc kubenswrapper[4894]: I1209 15:32:26.931057 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:26Z","lastTransitionTime":"2025-12-09T15:32:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.033863 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.033932 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.033957 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.033988 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.034013 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:27Z","lastTransitionTime":"2025-12-09T15:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.106179 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:27 crc kubenswrapper[4894]: E1209 15:32:27.106536 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.106290 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:27 crc kubenswrapper[4894]: E1209 15:32:27.106883 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.106220 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:27 crc kubenswrapper[4894]: E1209 15:32:27.107134 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.136949 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.136990 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.136998 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.137013 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.137038 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:27Z","lastTransitionTime":"2025-12-09T15:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.239832 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.239876 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.239890 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.239906 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.239917 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:27Z","lastTransitionTime":"2025-12-09T15:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.341838 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.341867 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.341877 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.341893 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.341903 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:27Z","lastTransitionTime":"2025-12-09T15:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.444292 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.444324 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.444332 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.444345 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.444355 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:27Z","lastTransitionTime":"2025-12-09T15:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.546825 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.546863 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.546874 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.546892 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.546905 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:27Z","lastTransitionTime":"2025-12-09T15:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.649799 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.649828 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.649840 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.649854 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.649862 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:27Z","lastTransitionTime":"2025-12-09T15:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.719436 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.719462 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.719470 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.719483 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.719492 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:27Z","lastTransitionTime":"2025-12-09T15:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:27 crc kubenswrapper[4894]: E1209 15:32:27.730059 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.733038 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.733075 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.733086 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.733102 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.733113 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:27Z","lastTransitionTime":"2025-12-09T15:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:27 crc kubenswrapper[4894]: E1209 15:32:27.744468 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.747620 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.747676 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.747689 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.747715 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.747728 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:27Z","lastTransitionTime":"2025-12-09T15:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:27 crc kubenswrapper[4894]: E1209 15:32:27.761132 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.765471 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.765510 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.765520 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.765538 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.765549 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:27Z","lastTransitionTime":"2025-12-09T15:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:27 crc kubenswrapper[4894]: E1209 15:32:27.777553 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.781005 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.781047 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.781056 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.781071 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.781082 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:27Z","lastTransitionTime":"2025-12-09T15:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:27 crc kubenswrapper[4894]: E1209 15:32:27.791535 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:27Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:27Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:27 crc kubenswrapper[4894]: E1209 15:32:27.791682 4894 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.793290 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.793361 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.793370 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.793382 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.793391 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:27Z","lastTransitionTime":"2025-12-09T15:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.895631 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.895711 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.895721 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.895752 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.895762 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:27Z","lastTransitionTime":"2025-12-09T15:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.997808 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.997868 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.997885 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.997910 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:27 crc kubenswrapper[4894]: I1209 15:32:27.997932 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:27Z","lastTransitionTime":"2025-12-09T15:32:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.100429 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.100482 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.100494 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.100513 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.100525 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:28Z","lastTransitionTime":"2025-12-09T15:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.106071 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:28 crc kubenswrapper[4894]: E1209 15:32:28.106218 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.203432 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.203480 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.203489 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.203504 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.203515 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:28Z","lastTransitionTime":"2025-12-09T15:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.306333 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.306376 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.306390 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.306406 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.306420 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:28Z","lastTransitionTime":"2025-12-09T15:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.408544 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.408583 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.408591 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.408606 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.408617 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:28Z","lastTransitionTime":"2025-12-09T15:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.511779 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.511851 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.511868 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.511893 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.511912 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:28Z","lastTransitionTime":"2025-12-09T15:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.615116 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.615157 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.615168 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.615184 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.615195 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:28Z","lastTransitionTime":"2025-12-09T15:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.718025 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.718081 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.718097 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.718115 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.718128 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:28Z","lastTransitionTime":"2025-12-09T15:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.820505 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.820550 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.820562 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.820580 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.820594 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:28Z","lastTransitionTime":"2025-12-09T15:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.923009 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.923052 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.923063 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.923079 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:28 crc kubenswrapper[4894]: I1209 15:32:28.923091 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:28Z","lastTransitionTime":"2025-12-09T15:32:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.026262 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.026318 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.026332 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.026351 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.026364 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:29Z","lastTransitionTime":"2025-12-09T15:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.106444 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.106483 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:29 crc kubenswrapper[4894]: E1209 15:32:29.106589 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.106519 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:29 crc kubenswrapper[4894]: E1209 15:32:29.106745 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:29 crc kubenswrapper[4894]: E1209 15:32:29.106950 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.128892 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.128930 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.128941 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.128959 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.128970 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:29Z","lastTransitionTime":"2025-12-09T15:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.231559 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.231609 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.231625 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.231668 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.231687 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:29Z","lastTransitionTime":"2025-12-09T15:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.334531 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.334583 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.334596 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.334614 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.334626 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:29Z","lastTransitionTime":"2025-12-09T15:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.436684 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.436735 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.436751 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.436774 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.436792 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:29Z","lastTransitionTime":"2025-12-09T15:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.538990 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.539048 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.539060 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.539088 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.539099 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:29Z","lastTransitionTime":"2025-12-09T15:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.641711 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.641760 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.641775 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.641792 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.641806 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:29Z","lastTransitionTime":"2025-12-09T15:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.744879 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.744916 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.744927 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.744944 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.744955 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:29Z","lastTransitionTime":"2025-12-09T15:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.847864 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.847939 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.847957 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.847981 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.848000 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:29Z","lastTransitionTime":"2025-12-09T15:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.950297 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.950365 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.950385 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.950409 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:29 crc kubenswrapper[4894]: I1209 15:32:29.950426 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:29Z","lastTransitionTime":"2025-12-09T15:32:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.053497 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.053538 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.053549 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.053565 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.053578 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:30Z","lastTransitionTime":"2025-12-09T15:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.106369 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:30 crc kubenswrapper[4894]: E1209 15:32:30.106556 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.156328 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.156375 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.156391 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.156409 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.156419 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:30Z","lastTransitionTime":"2025-12-09T15:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.258251 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.258286 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.258295 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.258308 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.258317 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:30Z","lastTransitionTime":"2025-12-09T15:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.360972 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.361053 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.361084 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.361118 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.361141 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:30Z","lastTransitionTime":"2025-12-09T15:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.464035 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.464096 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.464114 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.464139 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.464157 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:30Z","lastTransitionTime":"2025-12-09T15:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.566299 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.566332 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.566342 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.566357 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.566368 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:30Z","lastTransitionTime":"2025-12-09T15:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.668629 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.668770 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.668793 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.668822 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.668842 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:30Z","lastTransitionTime":"2025-12-09T15:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.771521 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.771558 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.771569 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.771585 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.771607 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:30Z","lastTransitionTime":"2025-12-09T15:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.874237 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.874304 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.874322 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.874347 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.874368 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:30Z","lastTransitionTime":"2025-12-09T15:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.977666 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.977793 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.977830 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.977869 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:30 crc kubenswrapper[4894]: I1209 15:32:30.977895 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:30Z","lastTransitionTime":"2025-12-09T15:32:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.080733 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.080799 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.080818 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.080844 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.080859 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:31Z","lastTransitionTime":"2025-12-09T15:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.106340 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.106387 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.106387 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:31 crc kubenswrapper[4894]: E1209 15:32:31.106481 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:31 crc kubenswrapper[4894]: E1209 15:32:31.106577 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:32:31 crc kubenswrapper[4894]: E1209 15:32:31.106701 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.183954 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.184037 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.184061 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.184092 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.184117 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:31Z","lastTransitionTime":"2025-12-09T15:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.287070 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.287120 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.287131 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.287148 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.287160 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:31Z","lastTransitionTime":"2025-12-09T15:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.393049 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.393088 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.393100 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.393125 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.393137 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:31Z","lastTransitionTime":"2025-12-09T15:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.495812 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.495849 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.495861 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.495878 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.495890 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:31Z","lastTransitionTime":"2025-12-09T15:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.598565 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.598610 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.598622 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.598656 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.598669 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:31Z","lastTransitionTime":"2025-12-09T15:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.702054 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.702160 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.702181 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.702770 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.703143 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:31Z","lastTransitionTime":"2025-12-09T15:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.805548 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.805585 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.805594 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.805610 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.805619 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:31Z","lastTransitionTime":"2025-12-09T15:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.907793 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.907834 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.907861 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.907875 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:31 crc kubenswrapper[4894]: I1209 15:32:31.907885 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:31Z","lastTransitionTime":"2025-12-09T15:32:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.010877 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.010960 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.010992 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.011025 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.011046 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:32Z","lastTransitionTime":"2025-12-09T15:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.096273 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.097500 4894 scope.go:117] "RemoveContainer" containerID="fbc99dc4587bee2241be7ff18f707d60bc3250e5c718d23a09884973cd22b24f" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.105497 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:32 crc kubenswrapper[4894]: E1209 15:32:32.105657 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.113541 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.113598 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.113616 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.113661 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.113677 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:32Z","lastTransitionTime":"2025-12-09T15:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.123788 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://fbc99dc4587bee2241be7ff18f707d60bc3250e5c718d23a09884973cd22b24f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc99dc4587bee2241be7ff18f707d60bc3250e5c718d23a09884973cd22b24f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"ift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1209 15:32:16.221102 6296 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 15:32:16.221118 6296 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1209 15:32:16.221130 6296 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1209 15:32:16.221193 6296 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-s5scr_openshift-ovn-kubernetes(97036c6f-dce3-444b-85e6-8476c16613c9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.136303 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.147920 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.160347 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.175119 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.186677 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.199068 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.212431 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.216002 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.216035 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.216047 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.216064 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.216077 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:32Z","lastTransitionTime":"2025-12-09T15:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.223723 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1473dc3d-be11-41e4-88d4-995a80be1e53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a183cf558c984ca981afe6cfbeda1cb173a511751d8edb975cf87e677e55c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://100782eeaa80dd7da85d32c80957a78582580bdf2bc44281f52b87865cffbe35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gsk62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.235782 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.249373 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qxfr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"948b28c5-e35c-4e6c-b31c-c0d5b205389b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qxfr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.263172 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.279552 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.293317 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.302029 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.312754 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.318744 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.318780 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.318788 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.318816 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.318826 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:32Z","lastTransitionTime":"2025-12-09T15:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.422561 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.422628 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.422674 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.422699 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.422717 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:32Z","lastTransitionTime":"2025-12-09T15:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.442227 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s5scr_97036c6f-dce3-444b-85e6-8476c16613c9/ovnkube-controller/1.log" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.445406 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerStarted","Data":"8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1"} Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.446125 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.462511 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.508378 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.523972 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.526090 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.526124 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.526179 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.526196 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.526206 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:32Z","lastTransitionTime":"2025-12-09T15:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.542492 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.560202 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qxfr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"948b28c5-e35c-4e6c-b31c-c0d5b205389b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qxfr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.577160 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.592068 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.614859 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.626973 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.628417 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.628518 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.628588 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.628703 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.628777 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:32Z","lastTransitionTime":"2025-12-09T15:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.642403 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.659943 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc99dc4587bee2241be7ff18f707d60bc3250e5c718d23a09884973cd22b24f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"ift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1209 15:32:16.221102 6296 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 15:32:16.221118 6296 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1209 15:32:16.221130 6296 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1209 15:32:16.221193 6296 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.672724 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.684725 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.696429 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.712460 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1473dc3d-be11-41e4-88d4-995a80be1e53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a183cf558c984ca981afe6cfbeda1cb173a511751d8edb975cf87e677e55c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://100782eeaa80dd7da85d32c80957a78582580bdf2bc44281f52b87865cffbe35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gsk62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.725290 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:32Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.730713 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.730749 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.730759 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.730774 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.730784 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:32Z","lastTransitionTime":"2025-12-09T15:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.833306 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.833354 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.833366 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.833380 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.833390 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:32Z","lastTransitionTime":"2025-12-09T15:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.936617 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.936693 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.936706 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.936728 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:32 crc kubenswrapper[4894]: I1209 15:32:32.936742 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:32Z","lastTransitionTime":"2025-12-09T15:32:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.039883 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.040311 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.040322 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.040341 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.040353 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:33Z","lastTransitionTime":"2025-12-09T15:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.105738 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.105777 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.105826 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:33 crc kubenswrapper[4894]: E1209 15:32:33.105878 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:32:33 crc kubenswrapper[4894]: E1209 15:32:33.105978 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:33 crc kubenswrapper[4894]: E1209 15:32:33.106065 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.142300 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.142344 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.142355 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.142371 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.142383 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:33Z","lastTransitionTime":"2025-12-09T15:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.245107 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.245157 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.245170 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.245214 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.245227 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:33Z","lastTransitionTime":"2025-12-09T15:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.347804 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.347869 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.347882 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.347904 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.347916 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:33Z","lastTransitionTime":"2025-12-09T15:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.450063 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.450150 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.450178 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.450211 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.450235 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:33Z","lastTransitionTime":"2025-12-09T15:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.451410 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s5scr_97036c6f-dce3-444b-85e6-8476c16613c9/ovnkube-controller/2.log" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.452212 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s5scr_97036c6f-dce3-444b-85e6-8476c16613c9/ovnkube-controller/1.log" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.465480 4894 generic.go:334] "Generic (PLEG): container finished" podID="97036c6f-dce3-444b-85e6-8476c16613c9" containerID="8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1" exitCode=1 Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.465577 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerDied","Data":"8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1"} Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.465627 4894 scope.go:117] "RemoveContainer" containerID="fbc99dc4587bee2241be7ff18f707d60bc3250e5c718d23a09884973cd22b24f" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.467353 4894 scope.go:117] "RemoveContainer" containerID="8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1" Dec 09 15:32:33 crc kubenswrapper[4894]: E1209 15:32:33.470177 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-s5scr_openshift-ovn-kubernetes(97036c6f-dce3-444b-85e6-8476c16613c9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.483042 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1473dc3d-be11-41e4-88d4-995a80be1e53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a183cf558c984ca981afe6cfbeda1cb173a511751d8edb975cf87e677e55c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://100782eeaa80dd7da85d32c80957a78582580bdf2bc44281f52b87865cffbe35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gsk62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:33Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.497784 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:33Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.510756 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:33Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.528394 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:33Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.540387 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:33Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.552894 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:33Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.553543 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.553588 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.553600 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.553617 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.553629 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:33Z","lastTransitionTime":"2025-12-09T15:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.566624 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qxfr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"948b28c5-e35c-4e6c-b31c-c0d5b205389b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qxfr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:33Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.593711 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:33Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.606560 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:33Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.622305 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:33Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.632943 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:33Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.648571 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:33Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.655488 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.655528 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.655537 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.655552 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.655560 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:33Z","lastTransitionTime":"2025-12-09T15:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.665936 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://fbc99dc4587bee2241be7ff18f707d60bc3250e5c718d23a09884973cd22b24f\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"message\\\":\\\"ift-kube-controller-manager/kube-controller-manager_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-kube-controller-manager/kube-controller-manager\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.4.36\\\\\\\", Port:443, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Templates:services.TemplateMap(nil), Switches:[]string{}, Routers:[]string{}, Groups:[]string{\\\\\\\"clusterLBGroup\\\\\\\"}}}\\\\nI1209 15:32:16.221102 6296 handler.go:208] Removed *v1.Node event handler 2\\\\nI1209 15:32:16.221118 6296 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1209 15:32:16.221130 6296 model_client.go:398] Mutate operations generated as: [{Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:73135118-cf1b-4568-bd31-2f50308bf69d}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nF1209 15:32:16.221193 6296 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:15Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:33Z\\\",\\\"message\\\":\\\"d3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 15:32:32.995178 6524 services_controller.go:451] Built service openshift-image-registry/image-registry cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.93\\\\\\\", Port:5000, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Temp\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:33Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.678286 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:33Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.690139 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:33Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.700836 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:33Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.757327 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.757356 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.757366 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.757378 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.757387 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:33Z","lastTransitionTime":"2025-12-09T15:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.859193 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.859231 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.859242 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.859258 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.859268 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:33Z","lastTransitionTime":"2025-12-09T15:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.961533 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.961571 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.961584 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.961600 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:33 crc kubenswrapper[4894]: I1209 15:32:33.961610 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:33Z","lastTransitionTime":"2025-12-09T15:32:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.065064 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.065138 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.065165 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.065226 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.065250 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:34Z","lastTransitionTime":"2025-12-09T15:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.105859 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:34 crc kubenswrapper[4894]: E1209 15:32:34.105996 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.168869 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.168937 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.168953 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.168989 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.169026 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:34Z","lastTransitionTime":"2025-12-09T15:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.271906 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.271969 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.271983 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.272005 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.272021 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:34Z","lastTransitionTime":"2025-12-09T15:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.374974 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.375021 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.375032 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.375049 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.375060 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:34Z","lastTransitionTime":"2025-12-09T15:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.469826 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s5scr_97036c6f-dce3-444b-85e6-8476c16613c9/ovnkube-controller/2.log" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.473509 4894 scope.go:117] "RemoveContainer" containerID="8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1" Dec 09 15:32:34 crc kubenswrapper[4894]: E1209 15:32:34.473669 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-s5scr_openshift-ovn-kubernetes(97036c6f-dce3-444b-85e6-8476c16613c9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.476580 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.476606 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.476615 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.476626 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.476655 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:34Z","lastTransitionTime":"2025-12-09T15:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.487585 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:34Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.501495 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:34Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.514611 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1473dc3d-be11-41e4-88d4-995a80be1e53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a183cf558c984ca981afe6cfbeda1cb173a511751d8edb975cf87e677e55c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://100782eeaa80dd7da85d32c80957a78582580bdf2bc44281f52b87865cffbe35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gsk62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:34Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.534245 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:34Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.546975 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:34Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.560913 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:34Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.570987 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:34Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.578887 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.578930 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.578946 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.578966 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.578983 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:34Z","lastTransitionTime":"2025-12-09T15:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.584176 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:34Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.595616 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qxfr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"948b28c5-e35c-4e6c-b31c-c0d5b205389b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qxfr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:34Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.623144 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:34Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.642836 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:34Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.655019 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:34Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.665521 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:34Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.681313 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.681371 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.681385 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.681405 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.681421 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:34Z","lastTransitionTime":"2025-12-09T15:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.681448 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:34Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.700103 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:33Z\\\",\\\"message\\\":\\\"d3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 15:32:32.995178 6524 services_controller.go:451] Built service openshift-image-registry/image-registry cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.93\\\\\\\", Port:5000, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Temp\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-s5scr_openshift-ovn-kubernetes(97036c6f-dce3-444b-85e6-8476c16613c9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:34Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.714046 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:34Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.779011 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs\") pod \"network-metrics-daemon-qxfr8\" (UID: \"948b28c5-e35c-4e6c-b31c-c0d5b205389b\") " pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:34 crc kubenswrapper[4894]: E1209 15:32:34.779190 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 15:32:34 crc kubenswrapper[4894]: E1209 15:32:34.779256 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs podName:948b28c5-e35c-4e6c-b31c-c0d5b205389b nodeName:}" failed. No retries permitted until 2025-12-09 15:32:50.779237354 +0000 UTC m=+65.098448023 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs") pod "network-metrics-daemon-qxfr8" (UID: "948b28c5-e35c-4e6c-b31c-c0d5b205389b") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.783223 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.783257 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.783268 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.783282 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.783291 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:34Z","lastTransitionTime":"2025-12-09T15:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.885979 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.886010 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.886019 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.886034 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.886046 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:34Z","lastTransitionTime":"2025-12-09T15:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.988568 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.988648 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.988660 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.988675 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:34 crc kubenswrapper[4894]: I1209 15:32:34.988686 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:34Z","lastTransitionTime":"2025-12-09T15:32:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.091049 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.091083 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.091095 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.091110 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.091121 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:35Z","lastTransitionTime":"2025-12-09T15:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.105818 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.105833 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:35 crc kubenswrapper[4894]: E1209 15:32:35.105943 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:35 crc kubenswrapper[4894]: E1209 15:32:35.106065 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.105839 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:35 crc kubenswrapper[4894]: E1209 15:32:35.106160 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.193443 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.193482 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.193491 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.193506 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.193515 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:35Z","lastTransitionTime":"2025-12-09T15:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.296589 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.296672 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.296687 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.296723 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.296734 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:35Z","lastTransitionTime":"2025-12-09T15:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.399042 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.399097 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.399109 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.399127 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.399142 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:35Z","lastTransitionTime":"2025-12-09T15:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.502417 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.502463 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.502473 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.502487 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.502496 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:35Z","lastTransitionTime":"2025-12-09T15:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.605273 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.605323 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.605336 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.605357 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.605372 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:35Z","lastTransitionTime":"2025-12-09T15:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.708191 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.708258 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.708277 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.708303 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.708323 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:35Z","lastTransitionTime":"2025-12-09T15:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.810419 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.810476 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.810485 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.810499 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.810508 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:35Z","lastTransitionTime":"2025-12-09T15:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.912897 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.912956 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.912971 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.912992 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:35 crc kubenswrapper[4894]: I1209 15:32:35.913006 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:35Z","lastTransitionTime":"2025-12-09T15:32:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.019994 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.020064 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.020084 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.020110 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.020133 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:36Z","lastTransitionTime":"2025-12-09T15:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.105955 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:36 crc kubenswrapper[4894]: E1209 15:32:36.106078 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.120284 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.122575 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.122678 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.122691 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.122705 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.122716 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:36Z","lastTransitionTime":"2025-12-09T15:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.139965 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.158686 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.174685 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.189767 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.204096 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qxfr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"948b28c5-e35c-4e6c-b31c-c0d5b205389b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qxfr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.212589 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.216740 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.227554 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.227597 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.227610 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.227628 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.227661 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:36Z","lastTransitionTime":"2025-12-09T15:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.227591 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.230100 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.243406 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.264019 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.281110 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:33Z\\\",\\\"message\\\":\\\"d3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 15:32:32.995178 6524 services_controller.go:451] Built service openshift-image-registry/image-registry cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.93\\\\\\\", Port:5000, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Temp\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-s5scr_openshift-ovn-kubernetes(97036c6f-dce3-444b-85e6-8476c16613c9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.293571 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.310856 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.330379 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.332211 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.332238 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.332416 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.332437 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.332449 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:36Z","lastTransitionTime":"2025-12-09T15:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.344446 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1473dc3d-be11-41e4-88d4-995a80be1e53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a183cf558c984ca981afe6cfbeda1cb173a511751d8edb975cf87e677e55c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://100782eeaa80dd7da85d32c80957a78582580bdf2bc44281f52b87865cffbe35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gsk62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.354703 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.365624 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.377821 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"376c4e26-e0f9-42cb-850d-7ffc8b29e099\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5293b95acee426113021da42be9937be7bcefa80361ac30667ee8dad75286dee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://882ae42ef0bd79cba2c436b8eeda122b33e1ed9efe5d09af3c80b3b1145c8103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b70791118fefd6a669a2e12536d1942d7120c7759a53e131f889242881918821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://771f15a0047608775ea7c49c6fd52f868423765c9ba57a43647c830947bb0ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://771f15a0047608775ea7c49c6fd52f868423765c9ba57a43647c830947bb0ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.390451 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.402374 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.413342 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.424525 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.436422 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qxfr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"948b28c5-e35c-4e6c-b31c-c0d5b205389b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qxfr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.436692 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.436784 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.436811 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.436873 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.436892 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:36Z","lastTransitionTime":"2025-12-09T15:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.450122 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.463761 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.479231 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.490524 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.504474 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.522482 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:33Z\\\",\\\"message\\\":\\\"d3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 15:32:32.995178 6524 services_controller.go:451] Built service openshift-image-registry/image-registry cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.93\\\\\\\", Port:5000, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Temp\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-s5scr_openshift-ovn-kubernetes(97036c6f-dce3-444b-85e6-8476c16613c9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.533902 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.539331 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.539356 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.539365 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.539378 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.539387 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:36Z","lastTransitionTime":"2025-12-09T15:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.544663 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.559498 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.570298 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1473dc3d-be11-41e4-88d4-995a80be1e53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a183cf558c984ca981afe6cfbeda1cb173a511751d8edb975cf87e677e55c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://100782eeaa80dd7da85d32c80957a78582580bdf2bc44281f52b87865cffbe35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gsk62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:36Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.642256 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.642311 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.642320 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.642349 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.642361 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:36Z","lastTransitionTime":"2025-12-09T15:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.745385 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.745420 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.745427 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.745458 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.745469 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:36Z","lastTransitionTime":"2025-12-09T15:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.848195 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.848251 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.848263 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.848282 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.848298 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:36Z","lastTransitionTime":"2025-12-09T15:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.951401 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.951472 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.951493 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.951520 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:36 crc kubenswrapper[4894]: I1209 15:32:36.951543 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:36Z","lastTransitionTime":"2025-12-09T15:32:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.002692 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.002900 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.002942 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:37 crc kubenswrapper[4894]: E1209 15:32:37.003072 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 15:32:37 crc kubenswrapper[4894]: E1209 15:32:37.003072 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:33:09.003014943 +0000 UTC m=+83.322225652 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:32:37 crc kubenswrapper[4894]: E1209 15:32:37.003093 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 15:32:37 crc kubenswrapper[4894]: E1209 15:32:37.003146 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 15:33:09.003122836 +0000 UTC m=+83.322333725 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 15:32:37 crc kubenswrapper[4894]: E1209 15:32:37.003232 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 15:33:09.003201718 +0000 UTC m=+83.322412387 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.054764 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.054822 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.054833 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.054854 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.054872 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:37Z","lastTransitionTime":"2025-12-09T15:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.104294 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.104361 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:37 crc kubenswrapper[4894]: E1209 15:32:37.104608 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 15:32:37 crc kubenswrapper[4894]: E1209 15:32:37.104689 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 15:32:37 crc kubenswrapper[4894]: E1209 15:32:37.104715 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:37 crc kubenswrapper[4894]: E1209 15:32:37.104753 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 15:32:37 crc kubenswrapper[4894]: E1209 15:32:37.104785 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 15:33:09.104767624 +0000 UTC m=+83.423978303 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:37 crc kubenswrapper[4894]: E1209 15:32:37.104792 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 15:32:37 crc kubenswrapper[4894]: E1209 15:32:37.104811 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:37 crc kubenswrapper[4894]: E1209 15:32:37.104904 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 15:33:09.104878267 +0000 UTC m=+83.424088936 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.106084 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.106115 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.106212 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:37 crc kubenswrapper[4894]: E1209 15:32:37.106319 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:37 crc kubenswrapper[4894]: E1209 15:32:37.106237 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:32:37 crc kubenswrapper[4894]: E1209 15:32:37.106497 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.158335 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.158383 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.158398 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.158419 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.158431 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:37Z","lastTransitionTime":"2025-12-09T15:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.260599 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.260678 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.260694 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.260716 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.260729 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:37Z","lastTransitionTime":"2025-12-09T15:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.365246 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.365351 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.365369 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.365401 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.365425 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:37Z","lastTransitionTime":"2025-12-09T15:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.468359 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.468409 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.468426 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.468444 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.468457 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:37Z","lastTransitionTime":"2025-12-09T15:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.571823 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.571896 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.571916 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.571942 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.571977 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:37Z","lastTransitionTime":"2025-12-09T15:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.674520 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.674671 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.674690 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.674715 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.674733 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:37Z","lastTransitionTime":"2025-12-09T15:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.777938 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.777980 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.778013 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.778031 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.778045 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:37Z","lastTransitionTime":"2025-12-09T15:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.880617 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.880673 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.880682 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.880696 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.880706 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:37Z","lastTransitionTime":"2025-12-09T15:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.983131 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.983177 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.983193 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.983209 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:37 crc kubenswrapper[4894]: I1209 15:32:37.983220 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:37Z","lastTransitionTime":"2025-12-09T15:32:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.016727 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.016766 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.016775 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.016790 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.016803 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:38Z","lastTransitionTime":"2025-12-09T15:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:38 crc kubenswrapper[4894]: E1209 15:32:38.033471 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.036989 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.037055 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.037068 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.037084 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.037095 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:38Z","lastTransitionTime":"2025-12-09T15:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:38 crc kubenswrapper[4894]: E1209 15:32:38.048178 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.051255 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.051280 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.051289 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.051304 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.051314 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:38Z","lastTransitionTime":"2025-12-09T15:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:38 crc kubenswrapper[4894]: E1209 15:32:38.061591 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.064851 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.064881 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.064890 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.064906 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.064915 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:38Z","lastTransitionTime":"2025-12-09T15:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:38 crc kubenswrapper[4894]: E1209 15:32:38.077787 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.081043 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.081075 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.081084 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.081101 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.081113 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:38Z","lastTransitionTime":"2025-12-09T15:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:38 crc kubenswrapper[4894]: E1209 15:32:38.093704 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:38Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:38 crc kubenswrapper[4894]: E1209 15:32:38.093841 4894 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.095588 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.095654 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.095665 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.095680 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.095691 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:38Z","lastTransitionTime":"2025-12-09T15:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.106122 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:38 crc kubenswrapper[4894]: E1209 15:32:38.106256 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.197629 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.197723 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.197741 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.197765 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.197786 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:38Z","lastTransitionTime":"2025-12-09T15:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.300197 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.300258 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.300283 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.300312 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.300336 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:38Z","lastTransitionTime":"2025-12-09T15:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.402310 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.402339 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.402349 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.402364 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.402374 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:38Z","lastTransitionTime":"2025-12-09T15:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.505056 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.505116 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.505133 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.505155 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.505172 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:38Z","lastTransitionTime":"2025-12-09T15:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.607563 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.607618 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.607631 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.607669 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.607682 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:38Z","lastTransitionTime":"2025-12-09T15:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.712081 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.712123 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.712134 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.712150 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.712164 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:38Z","lastTransitionTime":"2025-12-09T15:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.814874 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.814932 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.814949 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.815034 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.815053 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:38Z","lastTransitionTime":"2025-12-09T15:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.918043 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.918093 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.918104 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.918121 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:38 crc kubenswrapper[4894]: I1209 15:32:38.918132 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:38Z","lastTransitionTime":"2025-12-09T15:32:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.021228 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.021281 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.021292 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.021348 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.021364 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:39Z","lastTransitionTime":"2025-12-09T15:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.105466 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.105509 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.105563 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:39 crc kubenswrapper[4894]: E1209 15:32:39.105586 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:39 crc kubenswrapper[4894]: E1209 15:32:39.105654 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:39 crc kubenswrapper[4894]: E1209 15:32:39.105721 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.124033 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.124069 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.124104 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.124122 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.124132 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:39Z","lastTransitionTime":"2025-12-09T15:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.230627 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.230737 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.230756 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.230779 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.230797 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:39Z","lastTransitionTime":"2025-12-09T15:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.333905 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.333951 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.333976 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.334007 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.334021 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:39Z","lastTransitionTime":"2025-12-09T15:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.437200 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.437722 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.437842 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.437956 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.438040 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:39Z","lastTransitionTime":"2025-12-09T15:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.542681 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.542730 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.542750 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.542780 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.542801 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:39Z","lastTransitionTime":"2025-12-09T15:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.645713 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.645783 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.645800 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.645822 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.645841 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:39Z","lastTransitionTime":"2025-12-09T15:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.748101 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.748189 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.748213 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.748241 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.748261 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:39Z","lastTransitionTime":"2025-12-09T15:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.850339 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.850555 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.850795 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.850881 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.850973 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:39Z","lastTransitionTime":"2025-12-09T15:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.953307 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.953422 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.953446 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.953476 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:39 crc kubenswrapper[4894]: I1209 15:32:39.953499 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:39Z","lastTransitionTime":"2025-12-09T15:32:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.056652 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.056678 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.056686 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.056698 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.056707 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:40Z","lastTransitionTime":"2025-12-09T15:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.106131 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:40 crc kubenswrapper[4894]: E1209 15:32:40.106238 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.159322 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.159366 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.159374 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.159389 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.159398 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:40Z","lastTransitionTime":"2025-12-09T15:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.264354 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.264406 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.264423 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.264447 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.264464 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:40Z","lastTransitionTime":"2025-12-09T15:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.366857 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.366909 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.366926 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.366950 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.366968 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:40Z","lastTransitionTime":"2025-12-09T15:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.470152 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.470197 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.470207 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.470221 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.470231 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:40Z","lastTransitionTime":"2025-12-09T15:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.573041 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.573089 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.573101 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.573118 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.573130 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:40Z","lastTransitionTime":"2025-12-09T15:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.675156 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.675232 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.675251 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.675274 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.675291 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:40Z","lastTransitionTime":"2025-12-09T15:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.778404 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.778487 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.778507 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.778532 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.778551 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:40Z","lastTransitionTime":"2025-12-09T15:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.882065 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.882102 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.882112 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.882128 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.882138 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:40Z","lastTransitionTime":"2025-12-09T15:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.984596 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.984689 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.984708 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.984732 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:40 crc kubenswrapper[4894]: I1209 15:32:40.984750 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:40Z","lastTransitionTime":"2025-12-09T15:32:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.087502 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.087562 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.087579 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.087603 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.087621 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:41Z","lastTransitionTime":"2025-12-09T15:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.105361 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.105462 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:41 crc kubenswrapper[4894]: E1209 15:32:41.105610 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.105705 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:41 crc kubenswrapper[4894]: E1209 15:32:41.105859 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:41 crc kubenswrapper[4894]: E1209 15:32:41.106025 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.192098 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.192142 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.192153 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.192168 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.192177 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:41Z","lastTransitionTime":"2025-12-09T15:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.294268 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.294304 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.294316 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.294331 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.294341 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:41Z","lastTransitionTime":"2025-12-09T15:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.397534 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.397604 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.397685 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.397728 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.397750 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:41Z","lastTransitionTime":"2025-12-09T15:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.500136 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.500205 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.500216 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.500228 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.500238 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:41Z","lastTransitionTime":"2025-12-09T15:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.603134 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.603191 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.603209 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.603232 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.603248 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:41Z","lastTransitionTime":"2025-12-09T15:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.709767 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.709813 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.709829 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.709846 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.709856 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:41Z","lastTransitionTime":"2025-12-09T15:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.811484 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.811527 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.811536 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.811550 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.811560 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:41Z","lastTransitionTime":"2025-12-09T15:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.913613 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.913663 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.913682 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.913697 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:41 crc kubenswrapper[4894]: I1209 15:32:41.913706 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:41Z","lastTransitionTime":"2025-12-09T15:32:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.016331 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.016366 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.016375 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.016388 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.016397 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:42Z","lastTransitionTime":"2025-12-09T15:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.106418 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:42 crc kubenswrapper[4894]: E1209 15:32:42.106566 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.118813 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.118876 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.118899 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.118927 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.118951 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:42Z","lastTransitionTime":"2025-12-09T15:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.222247 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.222712 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.222896 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.223083 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.223291 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:42Z","lastTransitionTime":"2025-12-09T15:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.326782 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.327157 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.327442 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.327824 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.327897 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:42Z","lastTransitionTime":"2025-12-09T15:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.431008 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.431067 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.431081 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.431099 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.431112 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:42Z","lastTransitionTime":"2025-12-09T15:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.533124 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.533163 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.533172 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.533187 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.533198 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:42Z","lastTransitionTime":"2025-12-09T15:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.635819 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.635870 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.635885 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.635907 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.635921 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:42Z","lastTransitionTime":"2025-12-09T15:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.738330 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.738386 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.738408 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.738428 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.738441 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:42Z","lastTransitionTime":"2025-12-09T15:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.841520 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.841566 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.841577 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.841594 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.841606 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:42Z","lastTransitionTime":"2025-12-09T15:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.944158 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.944216 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.944233 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.944257 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:42 crc kubenswrapper[4894]: I1209 15:32:42.944276 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:42Z","lastTransitionTime":"2025-12-09T15:32:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.047351 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.047403 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.047421 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.047442 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.047457 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:43Z","lastTransitionTime":"2025-12-09T15:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.105989 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.105989 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.106124 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:43 crc kubenswrapper[4894]: E1209 15:32:43.106263 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:43 crc kubenswrapper[4894]: E1209 15:32:43.106431 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:43 crc kubenswrapper[4894]: E1209 15:32:43.106536 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.150932 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.151050 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.151073 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.151099 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.151116 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:43Z","lastTransitionTime":"2025-12-09T15:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.254732 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.254789 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.254802 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.254823 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.254836 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:43Z","lastTransitionTime":"2025-12-09T15:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.357413 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.357446 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.357453 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.357467 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.357474 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:43Z","lastTransitionTime":"2025-12-09T15:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.459953 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.459998 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.460011 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.460029 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.460040 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:43Z","lastTransitionTime":"2025-12-09T15:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.563728 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.563779 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.563800 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.563822 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.563837 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:43Z","lastTransitionTime":"2025-12-09T15:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.666506 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.666551 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.666560 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.666575 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.666584 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:43Z","lastTransitionTime":"2025-12-09T15:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.769721 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.769860 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.769877 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.769897 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.769910 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:43Z","lastTransitionTime":"2025-12-09T15:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.873363 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.873427 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.873440 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.873460 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.873472 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:43Z","lastTransitionTime":"2025-12-09T15:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.978037 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.978093 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.978109 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.978129 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:43 crc kubenswrapper[4894]: I1209 15:32:43.978144 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:43Z","lastTransitionTime":"2025-12-09T15:32:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.081294 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.081340 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.081350 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.081367 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.081378 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:44Z","lastTransitionTime":"2025-12-09T15:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.106490 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:44 crc kubenswrapper[4894]: E1209 15:32:44.106807 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.184158 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.184231 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.184247 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.184269 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.184282 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:44Z","lastTransitionTime":"2025-12-09T15:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.286004 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.286064 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.286075 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.286097 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.286109 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:44Z","lastTransitionTime":"2025-12-09T15:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.388567 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.388626 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.388666 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.388687 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.388699 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:44Z","lastTransitionTime":"2025-12-09T15:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.491755 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.491807 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.491817 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.491834 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.491845 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:44Z","lastTransitionTime":"2025-12-09T15:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.594882 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.594921 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.594933 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.594949 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.594958 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:44Z","lastTransitionTime":"2025-12-09T15:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.697268 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.697295 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.697305 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.697318 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.697326 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:44Z","lastTransitionTime":"2025-12-09T15:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.800670 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.800715 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.800727 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.800743 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.800759 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:44Z","lastTransitionTime":"2025-12-09T15:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.904457 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.904496 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.904510 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.904536 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:44 crc kubenswrapper[4894]: I1209 15:32:44.904549 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:44Z","lastTransitionTime":"2025-12-09T15:32:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.007157 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.007205 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.007215 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.007229 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.007239 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:45Z","lastTransitionTime":"2025-12-09T15:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.105456 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.105498 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.105480 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:45 crc kubenswrapper[4894]: E1209 15:32:45.105596 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:45 crc kubenswrapper[4894]: E1209 15:32:45.105775 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:45 crc kubenswrapper[4894]: E1209 15:32:45.105882 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.109046 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.109075 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.109103 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.109118 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.109127 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:45Z","lastTransitionTime":"2025-12-09T15:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.211922 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.212182 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.212251 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.212316 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.212375 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:45Z","lastTransitionTime":"2025-12-09T15:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.314382 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.314406 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.314414 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.314426 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.314435 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:45Z","lastTransitionTime":"2025-12-09T15:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.416817 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.416852 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.416862 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.416879 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.416890 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:45Z","lastTransitionTime":"2025-12-09T15:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.518729 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.518758 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.518766 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.518778 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.518788 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:45Z","lastTransitionTime":"2025-12-09T15:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.622394 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.622482 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.622504 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.622533 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.622554 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:45Z","lastTransitionTime":"2025-12-09T15:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.725885 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.725928 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.725936 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.725951 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.725962 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:45Z","lastTransitionTime":"2025-12-09T15:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.827966 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.828011 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.828023 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.828041 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.828052 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:45Z","lastTransitionTime":"2025-12-09T15:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.930685 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.930716 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.930723 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.930736 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:45 crc kubenswrapper[4894]: I1209 15:32:45.930751 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:45Z","lastTransitionTime":"2025-12-09T15:32:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.032354 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.032397 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.032411 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.032427 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.032436 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:46Z","lastTransitionTime":"2025-12-09T15:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.106993 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:46 crc kubenswrapper[4894]: E1209 15:32:46.107512 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.123655 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:46Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.135062 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.135127 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.135141 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.135163 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.135176 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:46Z","lastTransitionTime":"2025-12-09T15:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.143974 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:33Z\\\",\\\"message\\\":\\\"d3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 15:32:32.995178 6524 services_controller.go:451] Built service openshift-image-registry/image-registry cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.93\\\\\\\", Port:5000, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Temp\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-s5scr_openshift-ovn-kubernetes(97036c6f-dce3-444b-85e6-8476c16613c9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:46Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.156328 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:46Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.166319 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:46Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.176070 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:46Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.187120 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1473dc3d-be11-41e4-88d4-995a80be1e53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a183cf558c984ca981afe6cfbeda1cb173a511751d8edb975cf87e677e55c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://100782eeaa80dd7da85d32c80957a78582580bdf2bc44281f52b87865cffbe35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gsk62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:46Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.198895 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:46Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.208814 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:46Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.220750 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:46Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.230865 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"376c4e26-e0f9-42cb-850d-7ffc8b29e099\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5293b95acee426113021da42be9937be7bcefa80361ac30667ee8dad75286dee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://882ae42ef0bd79cba2c436b8eeda122b33e1ed9efe5d09af3c80b3b1145c8103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b70791118fefd6a669a2e12536d1942d7120c7759a53e131f889242881918821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://771f15a0047608775ea7c49c6fd52f868423765c9ba57a43647c830947bb0ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://771f15a0047608775ea7c49c6fd52f868423765c9ba57a43647c830947bb0ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:46Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.238231 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.238271 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.238280 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.238295 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.238304 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:46Z","lastTransitionTime":"2025-12-09T15:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.240205 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:46Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.250843 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:46Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.262460 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qxfr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"948b28c5-e35c-4e6c-b31c-c0d5b205389b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qxfr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:46Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.275539 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:46Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.287445 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:46Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.300159 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:46Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.309323 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:46Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.341660 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.341694 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.341705 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.341720 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.341732 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:46Z","lastTransitionTime":"2025-12-09T15:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.444093 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.444133 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.444144 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.444158 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.444168 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:46Z","lastTransitionTime":"2025-12-09T15:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.546501 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.546567 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.546588 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.546616 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.546637 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:46Z","lastTransitionTime":"2025-12-09T15:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.649556 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.649607 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.649621 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.649673 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.649690 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:46Z","lastTransitionTime":"2025-12-09T15:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.752381 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.752419 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.752429 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.752441 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.752450 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:46Z","lastTransitionTime":"2025-12-09T15:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.855774 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.855830 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.855849 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.855869 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.855885 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:46Z","lastTransitionTime":"2025-12-09T15:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.958141 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.958191 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.958201 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.958220 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:46 crc kubenswrapper[4894]: I1209 15:32:46.958231 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:46Z","lastTransitionTime":"2025-12-09T15:32:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.060375 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.060418 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.060431 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.060447 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.060458 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:47Z","lastTransitionTime":"2025-12-09T15:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.106000 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.106033 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.106022 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:47 crc kubenswrapper[4894]: E1209 15:32:47.106146 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:47 crc kubenswrapper[4894]: E1209 15:32:47.106450 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:32:47 crc kubenswrapper[4894]: E1209 15:32:47.106616 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.106835 4894 scope.go:117] "RemoveContainer" containerID="8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1" Dec 09 15:32:47 crc kubenswrapper[4894]: E1209 15:32:47.107003 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-s5scr_openshift-ovn-kubernetes(97036c6f-dce3-444b-85e6-8476c16613c9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.162326 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.162368 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.162380 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.162397 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.162410 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:47Z","lastTransitionTime":"2025-12-09T15:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.264592 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.264638 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.264677 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.264696 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.264708 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:47Z","lastTransitionTime":"2025-12-09T15:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.366953 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.367012 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.367025 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.367043 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.367054 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:47Z","lastTransitionTime":"2025-12-09T15:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.469638 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.469694 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.469706 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.469722 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.469733 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:47Z","lastTransitionTime":"2025-12-09T15:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.572043 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.572085 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.572100 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.572116 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.572128 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:47Z","lastTransitionTime":"2025-12-09T15:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.674022 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.674071 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.674083 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.674104 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.674121 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:47Z","lastTransitionTime":"2025-12-09T15:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.776950 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.776997 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.777007 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.777024 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.777040 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:47Z","lastTransitionTime":"2025-12-09T15:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.878956 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.879018 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.879036 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.879060 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.879079 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:47Z","lastTransitionTime":"2025-12-09T15:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.981410 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.981440 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.981450 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.981463 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:47 crc kubenswrapper[4894]: I1209 15:32:47.981471 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:47Z","lastTransitionTime":"2025-12-09T15:32:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.084152 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.084205 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.084223 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.084241 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.084254 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:48Z","lastTransitionTime":"2025-12-09T15:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.105907 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:48 crc kubenswrapper[4894]: E1209 15:32:48.106047 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.187294 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.187335 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.187345 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.187361 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.187374 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:48Z","lastTransitionTime":"2025-12-09T15:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.290405 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.290440 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.290450 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.290466 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.290475 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:48Z","lastTransitionTime":"2025-12-09T15:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.330658 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.330704 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.330714 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.330732 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.330742 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:48Z","lastTransitionTime":"2025-12-09T15:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:48 crc kubenswrapper[4894]: E1209 15:32:48.345550 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:48Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.348896 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.348917 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.348926 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.348938 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.348948 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:48Z","lastTransitionTime":"2025-12-09T15:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:48 crc kubenswrapper[4894]: E1209 15:32:48.364987 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:48Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.369710 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.369760 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.369774 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.369827 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.369847 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:48Z","lastTransitionTime":"2025-12-09T15:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:48 crc kubenswrapper[4894]: E1209 15:32:48.390083 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:48Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.394105 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.394142 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.394153 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.394170 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.394184 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:48Z","lastTransitionTime":"2025-12-09T15:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:48 crc kubenswrapper[4894]: E1209 15:32:48.409081 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:48Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.413101 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.413139 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.413152 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.413171 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.413186 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:48Z","lastTransitionTime":"2025-12-09T15:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:48 crc kubenswrapper[4894]: E1209 15:32:48.430872 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:48Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:48 crc kubenswrapper[4894]: E1209 15:32:48.431042 4894 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.432584 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.432647 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.432680 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.432697 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.432708 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:48Z","lastTransitionTime":"2025-12-09T15:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.535394 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.535422 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.535430 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.535444 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.535455 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:48Z","lastTransitionTime":"2025-12-09T15:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.638404 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.638454 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.638471 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.638518 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.638569 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:48Z","lastTransitionTime":"2025-12-09T15:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.741588 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.741626 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.741639 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.741685 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.741710 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:48Z","lastTransitionTime":"2025-12-09T15:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.844227 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.844267 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.844281 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.844298 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.844311 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:48Z","lastTransitionTime":"2025-12-09T15:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.946157 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.946200 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.946210 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.946226 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:48 crc kubenswrapper[4894]: I1209 15:32:48.946236 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:48Z","lastTransitionTime":"2025-12-09T15:32:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.048777 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.048810 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.048818 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.048831 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.048839 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:49Z","lastTransitionTime":"2025-12-09T15:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.105815 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.105850 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.105815 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:49 crc kubenswrapper[4894]: E1209 15:32:49.105957 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:49 crc kubenswrapper[4894]: E1209 15:32:49.106024 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:32:49 crc kubenswrapper[4894]: E1209 15:32:49.106088 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.151098 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.151144 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.151152 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.151166 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.151176 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:49Z","lastTransitionTime":"2025-12-09T15:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.252943 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.252995 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.253003 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.253017 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.253057 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:49Z","lastTransitionTime":"2025-12-09T15:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.355337 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.355378 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.355393 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.355410 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.355422 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:49Z","lastTransitionTime":"2025-12-09T15:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.458962 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.459017 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.459035 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.459057 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.459072 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:49Z","lastTransitionTime":"2025-12-09T15:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.561555 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.561601 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.561614 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.561630 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.561662 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:49Z","lastTransitionTime":"2025-12-09T15:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.663409 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.663448 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.663456 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.663471 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.663483 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:49Z","lastTransitionTime":"2025-12-09T15:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.765591 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.765803 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.765897 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.765983 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.766052 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:49Z","lastTransitionTime":"2025-12-09T15:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.868748 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.869005 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.869147 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.869260 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.869373 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:49Z","lastTransitionTime":"2025-12-09T15:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.971590 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.971853 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.971928 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.971994 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:49 crc kubenswrapper[4894]: I1209 15:32:49.972070 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:49Z","lastTransitionTime":"2025-12-09T15:32:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.073933 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.074328 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.074439 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.074779 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.075148 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:50Z","lastTransitionTime":"2025-12-09T15:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.106394 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:50 crc kubenswrapper[4894]: E1209 15:32:50.106525 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.177499 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.177846 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.177935 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.178019 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.178330 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:50Z","lastTransitionTime":"2025-12-09T15:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.280655 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.280720 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.280732 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.280749 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.280760 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:50Z","lastTransitionTime":"2025-12-09T15:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.383092 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.383143 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.383154 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.383171 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.383184 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:50Z","lastTransitionTime":"2025-12-09T15:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.485020 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.485055 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.485066 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.485083 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.485094 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:50Z","lastTransitionTime":"2025-12-09T15:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.587741 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.587785 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.587796 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.587813 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.587825 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:50Z","lastTransitionTime":"2025-12-09T15:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.689933 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.689970 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.689978 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.689991 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.690000 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:50Z","lastTransitionTime":"2025-12-09T15:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.791994 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.792056 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.792067 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.792099 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.792109 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:50Z","lastTransitionTime":"2025-12-09T15:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:50 crc kubenswrapper[4894]: E1209 15:32:50.850675 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 15:32:50 crc kubenswrapper[4894]: E1209 15:32:50.850750 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs podName:948b28c5-e35c-4e6c-b31c-c0d5b205389b nodeName:}" failed. No retries permitted until 2025-12-09 15:33:22.850728894 +0000 UTC m=+97.169939583 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs") pod "network-metrics-daemon-qxfr8" (UID: "948b28c5-e35c-4e6c-b31c-c0d5b205389b") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.850526 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs\") pod \"network-metrics-daemon-qxfr8\" (UID: \"948b28c5-e35c-4e6c-b31c-c0d5b205389b\") " pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.895024 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.895055 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.895064 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.895078 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.895088 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:50Z","lastTransitionTime":"2025-12-09T15:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.998020 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.998071 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.998083 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.998102 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:50 crc kubenswrapper[4894]: I1209 15:32:50.998115 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:50Z","lastTransitionTime":"2025-12-09T15:32:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.100583 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.100621 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.100631 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.100650 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.100676 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:51Z","lastTransitionTime":"2025-12-09T15:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.105827 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.105892 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:51 crc kubenswrapper[4894]: E1209 15:32:51.105936 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:51 crc kubenswrapper[4894]: E1209 15:32:51.106002 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.106064 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:51 crc kubenswrapper[4894]: E1209 15:32:51.106121 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.203167 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.203199 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.203208 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.203221 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.203230 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:51Z","lastTransitionTime":"2025-12-09T15:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.306117 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.306156 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.306167 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.306185 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.306194 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:51Z","lastTransitionTime":"2025-12-09T15:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.408029 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.408062 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.408070 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.408083 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.408091 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:51Z","lastTransitionTime":"2025-12-09T15:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.510246 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.510281 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.510291 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.510307 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.510318 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:51Z","lastTransitionTime":"2025-12-09T15:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.613146 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.613200 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.613215 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.613238 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.613249 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:51Z","lastTransitionTime":"2025-12-09T15:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.716052 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.716105 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.716113 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.716126 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.716136 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:51Z","lastTransitionTime":"2025-12-09T15:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.818522 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.818581 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.818598 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.818624 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.818670 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:51Z","lastTransitionTime":"2025-12-09T15:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.921455 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.921520 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.921533 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.921552 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:51 crc kubenswrapper[4894]: I1209 15:32:51.921565 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:51Z","lastTransitionTime":"2025-12-09T15:32:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.024003 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.024036 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.024045 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.024058 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.024069 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:52Z","lastTransitionTime":"2025-12-09T15:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.106067 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:52 crc kubenswrapper[4894]: E1209 15:32:52.106193 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.125593 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.125653 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.125663 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.125678 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.125690 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:52Z","lastTransitionTime":"2025-12-09T15:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.228144 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.228212 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.228231 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.228255 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.228272 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:52Z","lastTransitionTime":"2025-12-09T15:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.331091 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.331144 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.331157 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.331174 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.331186 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:52Z","lastTransitionTime":"2025-12-09T15:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.434220 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.434259 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.434269 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.434287 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.434298 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:52Z","lastTransitionTime":"2025-12-09T15:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.536406 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.536450 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.536461 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.536478 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.536491 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:52Z","lastTransitionTime":"2025-12-09T15:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.638842 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.638892 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.638904 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.638923 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.638935 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:52Z","lastTransitionTime":"2025-12-09T15:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.743169 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.743230 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.743244 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.743262 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.743280 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:52Z","lastTransitionTime":"2025-12-09T15:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.845700 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.845976 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.846063 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.846168 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.846267 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:52Z","lastTransitionTime":"2025-12-09T15:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.949370 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.949410 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.949418 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.949434 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:52 crc kubenswrapper[4894]: I1209 15:32:52.949443 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:52Z","lastTransitionTime":"2025-12-09T15:32:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.051857 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.051894 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.051904 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.051917 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.051927 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:53Z","lastTransitionTime":"2025-12-09T15:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.105924 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:53 crc kubenswrapper[4894]: E1209 15:32:53.106089 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.105928 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:53 crc kubenswrapper[4894]: E1209 15:32:53.106267 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.106735 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:53 crc kubenswrapper[4894]: E1209 15:32:53.107081 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.154544 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.154591 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.154599 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.154613 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.154625 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:53Z","lastTransitionTime":"2025-12-09T15:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.256983 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.257028 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.257041 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.257060 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.257072 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:53Z","lastTransitionTime":"2025-12-09T15:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.359504 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.360101 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.360206 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.360304 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.360400 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:53Z","lastTransitionTime":"2025-12-09T15:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.463021 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.463063 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.463075 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.463092 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.463104 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:53Z","lastTransitionTime":"2025-12-09T15:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.529223 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qt56s_97d690f0-ac29-46c9-ad4a-b92bd0a30772/kube-multus/0.log" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.529305 4894 generic.go:334] "Generic (PLEG): container finished" podID="97d690f0-ac29-46c9-ad4a-b92bd0a30772" containerID="51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e" exitCode=1 Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.529355 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qt56s" event={"ID":"97d690f0-ac29-46c9-ad4a-b92bd0a30772","Type":"ContainerDied","Data":"51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e"} Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.529986 4894 scope.go:117] "RemoveContainer" containerID="51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.542852 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:53Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.558520 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:53Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.570833 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:53Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.571785 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.571857 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.571877 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.571904 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.571921 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:53Z","lastTransitionTime":"2025-12-09T15:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.586215 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:53Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.600695 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qxfr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"948b28c5-e35c-4e6c-b31c-c0d5b205389b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qxfr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:53Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.618822 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:53Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.637675 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:53Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.657881 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:53Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.674200 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:53Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.675982 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.676037 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.676081 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.676109 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.676128 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:53Z","lastTransitionTime":"2025-12-09T15:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.696338 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:53Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.714153 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:33Z\\\",\\\"message\\\":\\\"d3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 15:32:32.995178 6524 services_controller.go:451] Built service openshift-image-registry/image-registry cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.93\\\\\\\", Port:5000, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Temp\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-s5scr_openshift-ovn-kubernetes(97036c6f-dce3-444b-85e6-8476c16613c9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:53Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.727344 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:53Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.742329 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:53Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.753956 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:53Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:52Z\\\",\\\"message\\\":\\\"2025-12-09T15:32:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_fa910caf-430c-4555-b9a3-e3f63500ddee\\\\n2025-12-09T15:32:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_fa910caf-430c-4555-b9a3-e3f63500ddee to /host/opt/cni/bin/\\\\n2025-12-09T15:32:06Z [verbose] multus-daemon started\\\\n2025-12-09T15:32:06Z [verbose] Readiness Indicator file check\\\\n2025-12-09T15:32:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:53Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.764950 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1473dc3d-be11-41e4-88d4-995a80be1e53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a183cf558c984ca981afe6cfbeda1cb173a511751d8edb975cf87e677e55c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://100782eeaa80dd7da85d32c80957a78582580bdf2bc44281f52b87865cffbe35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gsk62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:53Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.776415 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:53Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.778755 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.778791 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.778808 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.778825 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.778836 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:53Z","lastTransitionTime":"2025-12-09T15:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.788040 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"376c4e26-e0f9-42cb-850d-7ffc8b29e099\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5293b95acee426113021da42be9937be7bcefa80361ac30667ee8dad75286dee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://882ae42ef0bd79cba2c436b8eeda122b33e1ed9efe5d09af3c80b3b1145c8103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b70791118fefd6a669a2e12536d1942d7120c7759a53e131f889242881918821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://771f15a0047608775ea7c49c6fd52f868423765c9ba57a43647c830947bb0ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://771f15a0047608775ea7c49c6fd52f868423765c9ba57a43647c830947bb0ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:53Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.881760 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.881796 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.881804 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.881819 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.881828 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:53Z","lastTransitionTime":"2025-12-09T15:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.986340 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.986409 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.986420 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.986445 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:53 crc kubenswrapper[4894]: I1209 15:32:53.986461 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:53Z","lastTransitionTime":"2025-12-09T15:32:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.088789 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.088835 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.088848 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.088870 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.088887 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:54Z","lastTransitionTime":"2025-12-09T15:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.105413 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:54 crc kubenswrapper[4894]: E1209 15:32:54.106186 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.191334 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.191385 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.191397 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.191459 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.191479 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:54Z","lastTransitionTime":"2025-12-09T15:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.293821 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.293877 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.293892 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.293911 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.293923 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:54Z","lastTransitionTime":"2025-12-09T15:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.397135 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.397178 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.397188 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.397204 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.397214 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:54Z","lastTransitionTime":"2025-12-09T15:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.499981 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.500014 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.500022 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.500035 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.500044 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:54Z","lastTransitionTime":"2025-12-09T15:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.533215 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qt56s_97d690f0-ac29-46c9-ad4a-b92bd0a30772/kube-multus/0.log" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.533271 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qt56s" event={"ID":"97d690f0-ac29-46c9-ad4a-b92bd0a30772","Type":"ContainerStarted","Data":"0cfd57bf783aa0e666bbe98fa7fc1eb2cda69cbd538b72fbb95491efd23e2037"} Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.546684 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:54Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.558350 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cfd57bf783aa0e666bbe98fa7fc1eb2cda69cbd538b72fbb95491efd23e2037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:52Z\\\",\\\"message\\\":\\\"2025-12-09T15:32:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_fa910caf-430c-4555-b9a3-e3f63500ddee\\\\n2025-12-09T15:32:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_fa910caf-430c-4555-b9a3-e3f63500ddee to /host/opt/cni/bin/\\\\n2025-12-09T15:32:06Z [verbose] multus-daemon started\\\\n2025-12-09T15:32:06Z [verbose] Readiness Indicator file check\\\\n2025-12-09T15:32:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:54Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.568857 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1473dc3d-be11-41e4-88d4-995a80be1e53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a183cf558c984ca981afe6cfbeda1cb173a511751d8edb975cf87e677e55c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://100782eeaa80dd7da85d32c80957a78582580bdf2bc44281f52b87865cffbe35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gsk62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:54Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.583310 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:54Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.595605 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"376c4e26-e0f9-42cb-850d-7ffc8b29e099\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5293b95acee426113021da42be9937be7bcefa80361ac30667ee8dad75286dee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://882ae42ef0bd79cba2c436b8eeda122b33e1ed9efe5d09af3c80b3b1145c8103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b70791118fefd6a669a2e12536d1942d7120c7759a53e131f889242881918821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://771f15a0047608775ea7c49c6fd52f868423765c9ba57a43647c830947bb0ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://771f15a0047608775ea7c49c6fd52f868423765c9ba57a43647c830947bb0ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:54Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.602601 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.602637 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.602668 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.602705 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.602716 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:54Z","lastTransitionTime":"2025-12-09T15:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.609112 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:54Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.620992 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:54Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.630344 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:54Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.643184 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:54Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.655623 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qxfr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"948b28c5-e35c-4e6c-b31c-c0d5b205389b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qxfr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:54Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.670502 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:54Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.685116 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:54Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.698004 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:54Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.706043 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.706082 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.706101 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.706121 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.706133 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:54Z","lastTransitionTime":"2025-12-09T15:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.708961 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:54Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.722127 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:54Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.739875 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:33Z\\\",\\\"message\\\":\\\"d3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 15:32:32.995178 6524 services_controller.go:451] Built service openshift-image-registry/image-registry cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.93\\\\\\\", Port:5000, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Temp\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-s5scr_openshift-ovn-kubernetes(97036c6f-dce3-444b-85e6-8476c16613c9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:54Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.752429 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:54Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.807989 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.808030 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.808038 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.808051 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.808060 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:54Z","lastTransitionTime":"2025-12-09T15:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.910258 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.910291 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.910300 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.910312 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:54 crc kubenswrapper[4894]: I1209 15:32:54.910320 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:54Z","lastTransitionTime":"2025-12-09T15:32:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.012495 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.012520 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.012528 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.012541 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.012550 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:55Z","lastTransitionTime":"2025-12-09T15:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.106388 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.106390 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:55 crc kubenswrapper[4894]: E1209 15:32:55.106517 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:55 crc kubenswrapper[4894]: E1209 15:32:55.106595 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.106388 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:55 crc kubenswrapper[4894]: E1209 15:32:55.106721 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.115024 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.115068 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.115081 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.115099 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.115110 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:55Z","lastTransitionTime":"2025-12-09T15:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.217823 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.217881 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.217899 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.217922 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.217939 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:55Z","lastTransitionTime":"2025-12-09T15:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.320208 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.320268 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.320286 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.320321 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.320339 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:55Z","lastTransitionTime":"2025-12-09T15:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.424467 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.424511 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.424523 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.424541 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.424554 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:55Z","lastTransitionTime":"2025-12-09T15:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.526634 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.526696 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.526707 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.526725 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.526735 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:55Z","lastTransitionTime":"2025-12-09T15:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.629516 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.629565 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.629576 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.629594 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.629605 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:55Z","lastTransitionTime":"2025-12-09T15:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.732720 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.732776 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.732787 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.732806 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.732817 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:55Z","lastTransitionTime":"2025-12-09T15:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.836129 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.836195 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.836210 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.836236 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.836254 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:55Z","lastTransitionTime":"2025-12-09T15:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.938828 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.938896 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.938912 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.938930 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:55 crc kubenswrapper[4894]: I1209 15:32:55.938941 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:55Z","lastTransitionTime":"2025-12-09T15:32:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.041618 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.041671 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.041681 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.041695 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.041704 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:56Z","lastTransitionTime":"2025-12-09T15:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.105721 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:56 crc kubenswrapper[4894]: E1209 15:32:56.105863 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.121259 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:56Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.134344 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:56Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.143734 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.143771 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.143781 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.143798 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.143818 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:56Z","lastTransitionTime":"2025-12-09T15:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.145691 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:56Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.160629 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:56Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.179713 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:33Z\\\",\\\"message\\\":\\\"d3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 15:32:32.995178 6524 services_controller.go:451] Built service openshift-image-registry/image-registry cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.93\\\\\\\", Port:5000, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Temp\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-s5scr_openshift-ovn-kubernetes(97036c6f-dce3-444b-85e6-8476c16613c9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:56Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.191318 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:56Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.203409 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:56Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.217054 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cfd57bf783aa0e666bbe98fa7fc1eb2cda69cbd538b72fbb95491efd23e2037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:52Z\\\",\\\"message\\\":\\\"2025-12-09T15:32:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_fa910caf-430c-4555-b9a3-e3f63500ddee\\\\n2025-12-09T15:32:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_fa910caf-430c-4555-b9a3-e3f63500ddee to /host/opt/cni/bin/\\\\n2025-12-09T15:32:06Z [verbose] multus-daemon started\\\\n2025-12-09T15:32:06Z [verbose] Readiness Indicator file check\\\\n2025-12-09T15:32:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:56Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.227323 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1473dc3d-be11-41e4-88d4-995a80be1e53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a183cf558c984ca981afe6cfbeda1cb173a511751d8edb975cf87e677e55c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://100782eeaa80dd7da85d32c80957a78582580bdf2bc44281f52b87865cffbe35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gsk62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:56Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.239477 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:56Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.247241 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.247280 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.247291 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.247306 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.247315 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:56Z","lastTransitionTime":"2025-12-09T15:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.252297 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"376c4e26-e0f9-42cb-850d-7ffc8b29e099\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5293b95acee426113021da42be9937be7bcefa80361ac30667ee8dad75286dee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://882ae42ef0bd79cba2c436b8eeda122b33e1ed9efe5d09af3c80b3b1145c8103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b70791118fefd6a669a2e12536d1942d7120c7759a53e131f889242881918821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://771f15a0047608775ea7c49c6fd52f868423765c9ba57a43647c830947bb0ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://771f15a0047608775ea7c49c6fd52f868423765c9ba57a43647c830947bb0ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:56Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.267890 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:56Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.281175 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:56Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.293121 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:56Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.303978 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:56Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.314348 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qxfr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"948b28c5-e35c-4e6c-b31c-c0d5b205389b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qxfr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:56Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.328215 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:56Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.349885 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.349919 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.349929 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.349942 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.349952 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:56Z","lastTransitionTime":"2025-12-09T15:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.451725 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.451804 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.451814 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.451827 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.451838 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:56Z","lastTransitionTime":"2025-12-09T15:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.556983 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.557057 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.557078 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.557105 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.557130 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:56Z","lastTransitionTime":"2025-12-09T15:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.660532 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.660587 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.660600 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.660620 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.660632 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:56Z","lastTransitionTime":"2025-12-09T15:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.763177 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.763249 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.763272 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.763300 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.763322 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:56Z","lastTransitionTime":"2025-12-09T15:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.865543 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.865577 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.865589 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.865604 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.865616 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:56Z","lastTransitionTime":"2025-12-09T15:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.967928 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.967970 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.967982 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.968000 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:56 crc kubenswrapper[4894]: I1209 15:32:56.968015 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:56Z","lastTransitionTime":"2025-12-09T15:32:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.070719 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.070779 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.070791 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.070807 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.070842 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:57Z","lastTransitionTime":"2025-12-09T15:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.105698 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.105756 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:57 crc kubenswrapper[4894]: E1209 15:32:57.105849 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.105756 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:57 crc kubenswrapper[4894]: E1209 15:32:57.105950 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:57 crc kubenswrapper[4894]: E1209 15:32:57.106058 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.172765 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.172816 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.172831 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.172849 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.172863 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:57Z","lastTransitionTime":"2025-12-09T15:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.275732 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.275787 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.275800 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.275820 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.275834 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:57Z","lastTransitionTime":"2025-12-09T15:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.378728 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.378778 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.378792 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.378810 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.378823 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:57Z","lastTransitionTime":"2025-12-09T15:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.482209 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.482252 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.482262 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.482278 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.482290 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:57Z","lastTransitionTime":"2025-12-09T15:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.584894 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.584935 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.584943 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.584960 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.584970 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:57Z","lastTransitionTime":"2025-12-09T15:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.687759 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.687797 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.687810 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.687827 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.687837 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:57Z","lastTransitionTime":"2025-12-09T15:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.791020 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.791056 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.791065 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.791078 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.791089 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:57Z","lastTransitionTime":"2025-12-09T15:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.893526 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.893578 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.893588 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.893604 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.893612 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:57Z","lastTransitionTime":"2025-12-09T15:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.995862 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.995895 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.995903 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.995915 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:57 crc kubenswrapper[4894]: I1209 15:32:57.995924 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:57Z","lastTransitionTime":"2025-12-09T15:32:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.098452 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.098488 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.098496 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.098512 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.098532 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:58Z","lastTransitionTime":"2025-12-09T15:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.105876 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:32:58 crc kubenswrapper[4894]: E1209 15:32:58.106008 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.106583 4894 scope.go:117] "RemoveContainer" containerID="8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.201480 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.201500 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.201508 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.201520 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.201530 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:58Z","lastTransitionTime":"2025-12-09T15:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.304241 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.304283 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.304293 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.304309 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.304319 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:58Z","lastTransitionTime":"2025-12-09T15:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.406061 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.406097 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.406108 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.406126 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.406137 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:58Z","lastTransitionTime":"2025-12-09T15:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.507910 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.507946 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.507955 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.507969 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.507981 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:58Z","lastTransitionTime":"2025-12-09T15:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.544156 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s5scr_97036c6f-dce3-444b-85e6-8476c16613c9/ovnkube-controller/2.log" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.546359 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerStarted","Data":"876913572e03e323189c646a2ea4359c30b652499f20ab2c876828c0fcc5ee36"} Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.546784 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.559499 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:58Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.569980 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:58Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.581803 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cfd57bf783aa0e666bbe98fa7fc1eb2cda69cbd538b72fbb95491efd23e2037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:52Z\\\",\\\"message\\\":\\\"2025-12-09T15:32:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_fa910caf-430c-4555-b9a3-e3f63500ddee\\\\n2025-12-09T15:32:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_fa910caf-430c-4555-b9a3-e3f63500ddee to /host/opt/cni/bin/\\\\n2025-12-09T15:32:06Z [verbose] multus-daemon started\\\\n2025-12-09T15:32:06Z [verbose] Readiness Indicator file check\\\\n2025-12-09T15:32:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:58Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.596028 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1473dc3d-be11-41e4-88d4-995a80be1e53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a183cf558c984ca981afe6cfbeda1cb173a511751d8edb975cf87e677e55c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://100782eeaa80dd7da85d32c80957a78582580bdf2bc44281f52b87865cffbe35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gsk62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:58Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.608546 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"376c4e26-e0f9-42cb-850d-7ffc8b29e099\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5293b95acee426113021da42be9937be7bcefa80361ac30667ee8dad75286dee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://882ae42ef0bd79cba2c436b8eeda122b33e1ed9efe5d09af3c80b3b1145c8103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b70791118fefd6a669a2e12536d1942d7120c7759a53e131f889242881918821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://771f15a0047608775ea7c49c6fd52f868423765c9ba57a43647c830947bb0ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://771f15a0047608775ea7c49c6fd52f868423765c9ba57a43647c830947bb0ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:58Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.609996 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.610047 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.610060 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.610076 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.610086 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:58Z","lastTransitionTime":"2025-12-09T15:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.613659 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.613699 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.613708 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.613722 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.613731 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:58Z","lastTransitionTime":"2025-12-09T15:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.619108 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:58Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:58 crc kubenswrapper[4894]: E1209 15:32:58.626082 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:58Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.629988 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.630021 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.630029 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.630044 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.630053 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:58Z","lastTransitionTime":"2025-12-09T15:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.637149 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:58Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:58 crc kubenswrapper[4894]: E1209 15:32:58.640688 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:58Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.643886 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.643936 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.643945 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.643961 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.643973 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:58Z","lastTransitionTime":"2025-12-09T15:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.650881 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:58Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:58 crc kubenswrapper[4894]: E1209 15:32:58.656392 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:58Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.662482 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.662520 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.662529 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.662544 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.662553 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:58Z","lastTransitionTime":"2025-12-09T15:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.664197 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:58Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.677804 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:58Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:58 crc kubenswrapper[4894]: E1209 15:32:58.680058 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:58Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.682913 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.682946 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.682956 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.682969 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.682979 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:58Z","lastTransitionTime":"2025-12-09T15:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.692171 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:58Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:58 crc kubenswrapper[4894]: E1209 15:32:58.703856 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:58Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:58 crc kubenswrapper[4894]: E1209 15:32:58.704015 4894 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.707415 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qxfr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"948b28c5-e35c-4e6c-b31c-c0d5b205389b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qxfr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:58Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.712558 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.712588 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.712601 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.712618 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.712632 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:58Z","lastTransitionTime":"2025-12-09T15:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.720722 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:58Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.737265 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:58Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.753147 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:58Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.772237 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:58Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.795177 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://876913572e03e323189c646a2ea4359c30b652499f20ab2c876828c0fcc5ee36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:33Z\\\",\\\"message\\\":\\\"d3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 15:32:32.995178 6524 services_controller.go:451] Built service openshift-image-registry/image-registry cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.93\\\\\\\", Port:5000, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Temp\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:58Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.814868 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.814918 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.814930 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.814947 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.814958 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:58Z","lastTransitionTime":"2025-12-09T15:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.917481 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.917532 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.917545 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.917561 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:58 crc kubenswrapper[4894]: I1209 15:32:58.917574 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:58Z","lastTransitionTime":"2025-12-09T15:32:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.020362 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.020400 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.020410 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.020425 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.020434 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:59Z","lastTransitionTime":"2025-12-09T15:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.106224 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:32:59 crc kubenswrapper[4894]: E1209 15:32:59.106350 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.106554 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:32:59 crc kubenswrapper[4894]: E1209 15:32:59.106631 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.106795 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:32:59 crc kubenswrapper[4894]: E1209 15:32:59.106860 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.122745 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.122777 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.122788 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.122803 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.122815 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:59Z","lastTransitionTime":"2025-12-09T15:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.224989 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.225032 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.225042 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.225058 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.225070 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:59Z","lastTransitionTime":"2025-12-09T15:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.328442 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.328910 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.328925 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.328939 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.328951 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:59Z","lastTransitionTime":"2025-12-09T15:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.435059 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.435119 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.435131 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.435152 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.435165 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:59Z","lastTransitionTime":"2025-12-09T15:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.537729 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.537764 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.537774 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.537790 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.537801 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:59Z","lastTransitionTime":"2025-12-09T15:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.551006 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s5scr_97036c6f-dce3-444b-85e6-8476c16613c9/ovnkube-controller/3.log" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.551816 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s5scr_97036c6f-dce3-444b-85e6-8476c16613c9/ovnkube-controller/2.log" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.554378 4894 generic.go:334] "Generic (PLEG): container finished" podID="97036c6f-dce3-444b-85e6-8476c16613c9" containerID="876913572e03e323189c646a2ea4359c30b652499f20ab2c876828c0fcc5ee36" exitCode=1 Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.554422 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerDied","Data":"876913572e03e323189c646a2ea4359c30b652499f20ab2c876828c0fcc5ee36"} Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.554464 4894 scope.go:117] "RemoveContainer" containerID="8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.555106 4894 scope.go:117] "RemoveContainer" containerID="876913572e03e323189c646a2ea4359c30b652499f20ab2c876828c0fcc5ee36" Dec 09 15:32:59 crc kubenswrapper[4894]: E1209 15:32:59.555276 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-s5scr_openshift-ovn-kubernetes(97036c6f-dce3-444b-85e6-8476c16613c9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.570187 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:59Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.583263 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:59Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.597384 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:59Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.616012 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:59Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.637043 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://876913572e03e323189c646a2ea4359c30b652499f20ab2c876828c0fcc5ee36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8cedac8673c80844115f5a69ab782018f4d35bc1f8bc24f1f0a7ef495a8cc0d1\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:33Z\\\",\\\"message\\\":\\\"d3574b26}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:update Table:NAT Row:map[external_ip:192.168.126.11 logical_ip:10.217.0.4 options:{GoMap:map[stateless:false]} type:snat] Rows:[] Columns:[] Mutations:[] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {43933d5e-3c3b-4ff8-8926-04ac25de450e}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:} {Op:mutate Table:Logical_Router Row:map[] Rows:[] Columns:[] Mutations:[{Column:nat Mutator:insert Value:{GoSet:[{GoUUID:43933d5e-3c3b-4ff8-8926-04ac25de450e}]}}] Timeout:\\\\u003cnil\\\\u003e Where:[where column _uuid == {e3c4661a-36a6-47f0-a6c0-a4ee741f2224}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI1209 15:32:32.995178 6524 services_controller.go:451] Built service openshift-image-registry/image-registry cluster-wide LB for network=default: []services.LB{services.LB{Name:\\\\\\\"Service_openshift-image-registry/image-registry_TCP_cluster\\\\\\\", UUID:\\\\\\\"\\\\\\\", Protocol:\\\\\\\"TCP\\\\\\\", ExternalIDs:map[string]string{\\\\\\\"k8s.ovn.org/kind\\\\\\\":\\\\\\\"Service\\\\\\\", \\\\\\\"k8s.ovn.org/owner\\\\\\\":\\\\\\\"openshift-image-registry/image-registry\\\\\\\"}, Opts:services.LBOpts{Reject:true, EmptyLBEvents:false, AffinityTimeOut:0, SkipSNAT:false, Template:false, AddressFamily:\\\\\\\"\\\\\\\"}, Rules:[]services.LBRule{services.LBRule{Source:services.Addr{IP:\\\\\\\"10.217.5.93\\\\\\\", Port:5000, Template:(*services.Template)(nil)}, Targets:[]services.Addr{}}}, Temp\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:32Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://876913572e03e323189c646a2ea4359c30b652499f20ab2c876828c0fcc5ee36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:59Z\\\",\\\"message\\\":\\\"for service openshift-etcd/etcd for network=default\\\\nI1209 15:32:59.088142 6918 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-mw747 after 0 failed attempt(s)\\\\nI1209 15:32:59.088156 6918 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-mw747\\\\nI1209 15:32:59.088145 6918 services_controller.go:434] Service openshift-etcd/etcd retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{etcd openshift-etcd ad0a4b9d-2a7b-4f3f-9020-0c45d515459d 4800 0 2025-02-23 05:11:51 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:etcd] map[operator.openshift.io/spec-hash:0685cfaa0976bfb7ba58513629369c20bf05f4fba36949e982bdb43af328f0e1 prometheus.io/scheme:https prometheus.io/scrape:true service.alpha.openshift.io/serving-cert-secret-name:serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:etcd,Protocol:TCP,Port:2379,TargetPort:{0 2379 },NodePort:0,AppProtocol:nil,},ServicePort{Name:etcd-metrics,Protocol:TCP,Port:9979,TargetPort:{0 9979 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{etcd: true,},ClusterIP:10.217.5.253,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,Lo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:59Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.640960 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.640995 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.641008 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.641025 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.641040 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:59Z","lastTransitionTime":"2025-12-09T15:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.650849 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:59Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.663647 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:59Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.676319 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cfd57bf783aa0e666bbe98fa7fc1eb2cda69cbd538b72fbb95491efd23e2037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:52Z\\\",\\\"message\\\":\\\"2025-12-09T15:32:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_fa910caf-430c-4555-b9a3-e3f63500ddee\\\\n2025-12-09T15:32:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_fa910caf-430c-4555-b9a3-e3f63500ddee to /host/opt/cni/bin/\\\\n2025-12-09T15:32:06Z [verbose] multus-daemon started\\\\n2025-12-09T15:32:06Z [verbose] Readiness Indicator file check\\\\n2025-12-09T15:32:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:59Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.688483 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1473dc3d-be11-41e4-88d4-995a80be1e53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a183cf558c984ca981afe6cfbeda1cb173a511751d8edb975cf87e677e55c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://100782eeaa80dd7da85d32c80957a78582580bdf2bc44281f52b87865cffbe35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gsk62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:59Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.700694 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"376c4e26-e0f9-42cb-850d-7ffc8b29e099\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5293b95acee426113021da42be9937be7bcefa80361ac30667ee8dad75286dee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://882ae42ef0bd79cba2c436b8eeda122b33e1ed9efe5d09af3c80b3b1145c8103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b70791118fefd6a669a2e12536d1942d7120c7759a53e131f889242881918821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://771f15a0047608775ea7c49c6fd52f868423765c9ba57a43647c830947bb0ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://771f15a0047608775ea7c49c6fd52f868423765c9ba57a43647c830947bb0ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:59Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.713193 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:59Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.729282 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:59Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.743692 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.743721 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.743729 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.743741 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.743750 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:59Z","lastTransitionTime":"2025-12-09T15:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.757447 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:59Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.777478 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:59Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.792455 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:59Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.804613 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:59Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.815680 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qxfr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"948b28c5-e35c-4e6c-b31c-c0d5b205389b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qxfr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:32:59Z is after 2025-08-24T17:21:41Z" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.845964 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.845996 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.846006 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.846041 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.846056 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:59Z","lastTransitionTime":"2025-12-09T15:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.948162 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.948204 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.948216 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.948232 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:32:59 crc kubenswrapper[4894]: I1209 15:32:59.948243 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:32:59Z","lastTransitionTime":"2025-12-09T15:32:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.050004 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.050042 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.050051 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.050064 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.050072 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:00Z","lastTransitionTime":"2025-12-09T15:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.105900 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:00 crc kubenswrapper[4894]: E1209 15:33:00.106099 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.118941 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.152068 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.152121 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.152130 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.152145 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.152158 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:00Z","lastTransitionTime":"2025-12-09T15:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.254235 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.254271 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.254281 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.254294 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.254304 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:00Z","lastTransitionTime":"2025-12-09T15:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.356986 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.357018 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.357028 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.357042 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.357050 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:00Z","lastTransitionTime":"2025-12-09T15:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.459256 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.459294 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.459303 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.459320 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.459330 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:00Z","lastTransitionTime":"2025-12-09T15:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.560402 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s5scr_97036c6f-dce3-444b-85e6-8476c16613c9/ovnkube-controller/3.log" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.562949 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.562986 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.563000 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.563025 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.563042 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:00Z","lastTransitionTime":"2025-12-09T15:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.564526 4894 scope.go:117] "RemoveContainer" containerID="876913572e03e323189c646a2ea4359c30b652499f20ab2c876828c0fcc5ee36" Dec 09 15:33:00 crc kubenswrapper[4894]: E1209 15:33:00.565242 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-s5scr_openshift-ovn-kubernetes(97036c6f-dce3-444b-85e6-8476c16613c9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.581068 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:00Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.595435 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:00Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.608509 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:00Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.623978 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:00Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.644680 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://876913572e03e323189c646a2ea4359c30b652499f20ab2c876828c0fcc5ee36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://876913572e03e323189c646a2ea4359c30b652499f20ab2c876828c0fcc5ee36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:59Z\\\",\\\"message\\\":\\\"for service openshift-etcd/etcd for network=default\\\\nI1209 15:32:59.088142 6918 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-mw747 after 0 failed attempt(s)\\\\nI1209 15:32:59.088156 6918 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-mw747\\\\nI1209 15:32:59.088145 6918 services_controller.go:434] Service openshift-etcd/etcd retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{etcd openshift-etcd ad0a4b9d-2a7b-4f3f-9020-0c45d515459d 4800 0 2025-02-23 05:11:51 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:etcd] map[operator.openshift.io/spec-hash:0685cfaa0976bfb7ba58513629369c20bf05f4fba36949e982bdb43af328f0e1 prometheus.io/scheme:https prometheus.io/scrape:true service.alpha.openshift.io/serving-cert-secret-name:serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:etcd,Protocol:TCP,Port:2379,TargetPort:{0 2379 },NodePort:0,AppProtocol:nil,},ServicePort{Name:etcd-metrics,Protocol:TCP,Port:9979,TargetPort:{0 9979 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{etcd: true,},ClusterIP:10.217.5.253,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,Lo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-s5scr_openshift-ovn-kubernetes(97036c6f-dce3-444b-85e6-8476c16613c9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:00Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.656561 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41dec44f-071d-4469-bd77-c9645f1c880a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd9840b18660dd5f9a52708ce51bc209f5dd295b3b830019c0f9b20f6ec63618\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://210a487b2bd321692484f3a44391086b4062bff9e34d2391f10d9f83a6ee4696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a487b2bd321692484f3a44391086b4062bff9e34d2391f10d9f83a6ee4696\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:00Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.665283 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.665318 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.665328 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.665342 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.665351 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:00Z","lastTransitionTime":"2025-12-09T15:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.671204 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:00Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.681833 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:00Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.692928 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cfd57bf783aa0e666bbe98fa7fc1eb2cda69cbd538b72fbb95491efd23e2037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:52Z\\\",\\\"message\\\":\\\"2025-12-09T15:32:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_fa910caf-430c-4555-b9a3-e3f63500ddee\\\\n2025-12-09T15:32:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_fa910caf-430c-4555-b9a3-e3f63500ddee to /host/opt/cni/bin/\\\\n2025-12-09T15:32:06Z [verbose] multus-daemon started\\\\n2025-12-09T15:32:06Z [verbose] Readiness Indicator file check\\\\n2025-12-09T15:32:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:00Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.701838 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1473dc3d-be11-41e4-88d4-995a80be1e53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a183cf558c984ca981afe6cfbeda1cb173a511751d8edb975cf87e677e55c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://100782eeaa80dd7da85d32c80957a78582580bdf2bc44281f52b87865cffbe35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gsk62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:00Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.712490 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:00Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.723550 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"376c4e26-e0f9-42cb-850d-7ffc8b29e099\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5293b95acee426113021da42be9937be7bcefa80361ac30667ee8dad75286dee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://882ae42ef0bd79cba2c436b8eeda122b33e1ed9efe5d09af3c80b3b1145c8103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b70791118fefd6a669a2e12536d1942d7120c7759a53e131f889242881918821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://771f15a0047608775ea7c49c6fd52f868423765c9ba57a43647c830947bb0ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://771f15a0047608775ea7c49c6fd52f868423765c9ba57a43647c830947bb0ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:00Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.737377 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:00Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.751513 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:00Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.764438 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:00Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.768048 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.768078 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.768088 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.768101 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.768111 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:00Z","lastTransitionTime":"2025-12-09T15:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.774777 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:00Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.783926 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qxfr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"948b28c5-e35c-4e6c-b31c-c0d5b205389b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qxfr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:00Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.796674 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:00Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.871248 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.871310 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.871322 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.871341 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.871352 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:00Z","lastTransitionTime":"2025-12-09T15:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.973608 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.973677 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.973690 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.973705 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:00 crc kubenswrapper[4894]: I1209 15:33:00.973716 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:00Z","lastTransitionTime":"2025-12-09T15:33:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.075674 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.075705 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.075713 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.075726 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.075735 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:01Z","lastTransitionTime":"2025-12-09T15:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.106438 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.106517 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:01 crc kubenswrapper[4894]: E1209 15:33:01.106591 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.106448 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:01 crc kubenswrapper[4894]: E1209 15:33:01.106764 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:01 crc kubenswrapper[4894]: E1209 15:33:01.106930 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.178670 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.178713 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.178721 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.178735 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.178744 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:01Z","lastTransitionTime":"2025-12-09T15:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.281432 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.281467 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.281476 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.281489 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.281497 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:01Z","lastTransitionTime":"2025-12-09T15:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.384457 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.384505 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.384516 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.384531 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.384540 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:01Z","lastTransitionTime":"2025-12-09T15:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.486823 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.486860 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.486892 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.486908 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.486918 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:01Z","lastTransitionTime":"2025-12-09T15:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.590231 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.590290 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.590308 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.590327 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.590342 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:01Z","lastTransitionTime":"2025-12-09T15:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.693135 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.693186 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.693203 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.693223 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.693240 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:01Z","lastTransitionTime":"2025-12-09T15:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.797024 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.797064 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.797076 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.797096 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.797164 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:01Z","lastTransitionTime":"2025-12-09T15:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.900155 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.900231 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.900244 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.900264 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:01 crc kubenswrapper[4894]: I1209 15:33:01.900277 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:01Z","lastTransitionTime":"2025-12-09T15:33:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.003360 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.003408 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.003420 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.003437 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.003452 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:02Z","lastTransitionTime":"2025-12-09T15:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.105686 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:02 crc kubenswrapper[4894]: E1209 15:33:02.105850 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.106004 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.106022 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.106032 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.106044 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.106055 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:02Z","lastTransitionTime":"2025-12-09T15:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.208403 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.208435 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.208443 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.208456 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.208467 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:02Z","lastTransitionTime":"2025-12-09T15:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.310569 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.310600 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.310607 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.310619 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.310628 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:02Z","lastTransitionTime":"2025-12-09T15:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.412718 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.412762 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.412774 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.412792 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.412803 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:02Z","lastTransitionTime":"2025-12-09T15:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.515192 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.515234 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.515245 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.515260 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.515271 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:02Z","lastTransitionTime":"2025-12-09T15:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.618022 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.618055 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.618063 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.618075 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.618085 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:02Z","lastTransitionTime":"2025-12-09T15:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.720316 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.720364 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.720378 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.720395 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.720407 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:02Z","lastTransitionTime":"2025-12-09T15:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.822488 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.822536 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.822547 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.822565 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.822579 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:02Z","lastTransitionTime":"2025-12-09T15:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.925029 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.925079 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.925090 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.925107 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:02 crc kubenswrapper[4894]: I1209 15:33:02.925119 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:02Z","lastTransitionTime":"2025-12-09T15:33:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.027200 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.027235 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.027246 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.027262 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.027273 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:03Z","lastTransitionTime":"2025-12-09T15:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.106425 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.106511 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:03 crc kubenswrapper[4894]: E1209 15:33:03.106574 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.106426 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:03 crc kubenswrapper[4894]: E1209 15:33:03.106703 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:03 crc kubenswrapper[4894]: E1209 15:33:03.107023 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.130144 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.130209 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.130225 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.130247 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.130263 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:03Z","lastTransitionTime":"2025-12-09T15:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.232246 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.232292 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.232303 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.232318 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.232329 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:03Z","lastTransitionTime":"2025-12-09T15:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.334234 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.334278 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.334294 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.334311 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.334322 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:03Z","lastTransitionTime":"2025-12-09T15:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.436719 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.436768 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.436781 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.436796 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.436809 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:03Z","lastTransitionTime":"2025-12-09T15:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.539426 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.539476 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.539486 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.539503 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.539513 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:03Z","lastTransitionTime":"2025-12-09T15:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.641567 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.641617 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.641630 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.641687 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.641701 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:03Z","lastTransitionTime":"2025-12-09T15:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.743795 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.743842 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.743856 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.743874 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.743886 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:03Z","lastTransitionTime":"2025-12-09T15:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.846089 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.846190 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.846215 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.846249 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.846276 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:03Z","lastTransitionTime":"2025-12-09T15:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.948569 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.948880 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.948958 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.949033 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:03 crc kubenswrapper[4894]: I1209 15:33:03.949089 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:03Z","lastTransitionTime":"2025-12-09T15:33:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.052197 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.052250 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.052267 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.052288 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.052304 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:04Z","lastTransitionTime":"2025-12-09T15:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.106283 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:04 crc kubenswrapper[4894]: E1209 15:33:04.106478 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.154432 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.154491 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.154508 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.154531 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.154550 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:04Z","lastTransitionTime":"2025-12-09T15:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.256989 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.257032 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.257041 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.257057 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.257067 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:04Z","lastTransitionTime":"2025-12-09T15:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.359438 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.359482 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.359492 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.359507 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.359517 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:04Z","lastTransitionTime":"2025-12-09T15:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.462059 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.462119 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.462132 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.462149 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.462162 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:04Z","lastTransitionTime":"2025-12-09T15:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.568043 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.568115 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.568132 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.568159 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.568184 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:04Z","lastTransitionTime":"2025-12-09T15:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.670986 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.671035 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.671047 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.671067 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.671082 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:04Z","lastTransitionTime":"2025-12-09T15:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.773039 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.773078 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.773088 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.773102 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.773111 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:04Z","lastTransitionTime":"2025-12-09T15:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.875277 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.875522 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.875751 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.875883 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.875950 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:04Z","lastTransitionTime":"2025-12-09T15:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.978585 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.978619 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.978697 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.978717 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:04 crc kubenswrapper[4894]: I1209 15:33:04.978727 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:04Z","lastTransitionTime":"2025-12-09T15:33:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.081324 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.081386 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.081400 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.081416 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.081428 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:05Z","lastTransitionTime":"2025-12-09T15:33:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.106190 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.106283 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.106190 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:05 crc kubenswrapper[4894]: E1209 15:33:05.106338 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:05 crc kubenswrapper[4894]: E1209 15:33:05.106420 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:05 crc kubenswrapper[4894]: E1209 15:33:05.106529 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.183937 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.183996 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.184007 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.184023 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.184035 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:05Z","lastTransitionTime":"2025-12-09T15:33:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.286176 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.286220 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.286229 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.286245 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.286254 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:05Z","lastTransitionTime":"2025-12-09T15:33:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.388302 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.388355 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.388369 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.388387 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.388399 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:05Z","lastTransitionTime":"2025-12-09T15:33:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.490662 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.490709 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.490723 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.490740 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.490752 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:05Z","lastTransitionTime":"2025-12-09T15:33:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.592678 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.592756 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.592774 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.592798 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.592815 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:05Z","lastTransitionTime":"2025-12-09T15:33:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.695887 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.695933 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.695946 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.695962 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.695975 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:05Z","lastTransitionTime":"2025-12-09T15:33:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.798425 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.798474 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.798483 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.798500 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.798509 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:05Z","lastTransitionTime":"2025-12-09T15:33:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.901277 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.901327 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.901336 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.901350 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:05 crc kubenswrapper[4894]: I1209 15:33:05.901358 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:05Z","lastTransitionTime":"2025-12-09T15:33:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.003782 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.003822 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.003837 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.003852 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.003863 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:06Z","lastTransitionTime":"2025-12-09T15:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.105403 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:06 crc kubenswrapper[4894]: E1209 15:33:06.105749 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.107016 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.107050 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.107060 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.107087 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.107097 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:06Z","lastTransitionTime":"2025-12-09T15:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.122459 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"98b81cce-6d1e-4537-822f-eb233e7a77f5\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4dbbbfe3f96f82207b45c667083f5cd235f761bcb41e7c7cb13c5d978e4cb85\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ce63abe0f425ba8188c101d549e7222527cbab2802a1dfb996e4ee9332f3115\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://93264af04a8c8e5ac2f41ee05ee951622c937006da962af0245490a817b89119\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.136750 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.147881 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-mw747" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"4b263411-430a-4950-8e4d-b0f305874889\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4e6cf4c5ca4bbb28bbc1dc9f2a689e5ea26f40efce5d365f94b9043cef610bb1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-f4wws\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-mw747\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.159511 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"cfa97fcb-7e7c-4202-a2d0-d283fa49b112\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6045a9ae4d546153e4d65be8394c4cc1dca4dd8db25e775925c1e8bfc65593dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-phrks\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-rj7cb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.168964 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-qxfr8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"948b28c5-e35c-4e6c-b31c-c0d5b205389b\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:18Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bggtq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:18Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-qxfr8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.180452 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9b8358d3-529c-43c9-964c-fe3bae20f1f8\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:24Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-12-09T15:32:04Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1209 15:31:58.550218 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1209 15:31:58.552452 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3185878112/tls.crt::/tmp/serving-cert-3185878112/tls.key\\\\\\\"\\\\nI1209 15:32:04.221573 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1209 15:32:04.224133 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1209 15:32:04.224152 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1209 15:32:04.224172 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1209 15:32:04.224177 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1209 15:32:04.230998 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1209 15:32:04.231019 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nI1209 15:32:04.231018 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1209 15:32:04.231024 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1209 15:32:04.231045 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1209 15:32:04.231048 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1209 15:32:04.231052 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1209 15:32:04.231055 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF1209 15:32:04.237362 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:48Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.192689 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.208068 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://2e1c8e45df2b3b573da161cedca99263cb3eb2a15922f9eee74dcb4281025791\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c3687af3e9d647b19c201bc280bf9bccda795012a07d29acf253a1d758b5ac13\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.209150 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.209185 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.209193 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.209206 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.209214 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:06Z","lastTransitionTime":"2025-12-09T15:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.220323 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8564888edad9b5c2f6ba54a2781382f7ebc88629d6d5349930cdde08312107f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.235339 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-54wfs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"05e6de1b-a262-4d89-ad3b-902a436d44cc\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:13Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6425d7a5123153dd0cb6448e47256d4654df9617936a27d657ed519f393e62b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3ee74729dd1cc936de1395cbd8ab215ffad057072397386f181170e09016fa82\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://ff38bee875f2543b89dca3749a6e03989bc021a385a14feedc3703bf7d4870e7\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://075d06d9f418f82658bab6e5a7d53505eccb0ec6e1a8ad4291d6ce87eff7f017\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e82ba7f82e5437a109cf866a923f35a881ec62a91cde9c9a19ec8599ac7da792\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:09Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:09Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://355ad2099d76794f35225990143338615e715a703a16ce807c97127383ac8317\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:10Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://880f5b61b8db0733b3541a1054bee8825171a818ee42cd7c678650a927f17f6f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:11Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:11Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-5jm69\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-54wfs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.253523 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97036c6f-dce3-444b-85e6-8476c16613c9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://876913572e03e323189c646a2ea4359c30b652499f20ab2c876828c0fcc5ee36\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://876913572e03e323189c646a2ea4359c30b652499f20ab2c876828c0fcc5ee36\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:59Z\\\",\\\"message\\\":\\\"for service openshift-etcd/etcd for network=default\\\\nI1209 15:32:59.088142 6918 obj_retry.go:386] Retry successful for *v1.Pod openshift-dns/node-resolver-mw747 after 0 failed attempt(s)\\\\nI1209 15:32:59.088156 6918 default_network_controller.go:776] Recording success event on pod openshift-dns/node-resolver-mw747\\\\nI1209 15:32:59.088145 6918 services_controller.go:434] Service openshift-etcd/etcd retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{etcd openshift-etcd ad0a4b9d-2a7b-4f3f-9020-0c45d515459d 4800 0 2025-02-23 05:11:51 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[k8s-app:etcd] map[operator.openshift.io/spec-hash:0685cfaa0976bfb7ba58513629369c20bf05f4fba36949e982bdb43af328f0e1 prometheus.io/scheme:https prometheus.io/scrape:true service.alpha.openshift.io/serving-cert-secret-name:serving-cert service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [] [] []},Spec:ServiceSpec{Ports:[]ServicePort{ServicePort{Name:etcd,Protocol:TCP,Port:2379,TargetPort:{0 2379 },NodePort:0,AppProtocol:nil,},ServicePort{Name:etcd-metrics,Protocol:TCP,Port:9979,TargetPort:{0 9979 },NodePort:0,AppProtocol:nil,},},Selector:map[string]string{etcd: true,},ClusterIP:10.217.5.253,Type:ClusterIP,ExternalIPs:[],SessionAffinity:None,Lo\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:58Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-s5scr_openshift-ovn-kubernetes(97036c6f-dce3-444b-85e6-8476c16613c9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-d5l7x\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-s5scr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.262866 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"41dec44f-071d-4469-bd77-c9645f1c880a\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:48Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://fd9840b18660dd5f9a52708ce51bc209f5dd295b3b830019c0f9b20f6ec63618\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://210a487b2bd321692484f3a44391086b4062bff9e34d2391f10d9f83a6ee4696\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://210a487b2bd321692484f3a44391086b4062bff9e34d2391f10d9f83a6ee4696\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.277529 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8e79bc256422296fe9ef22e598ec2f6e91d427521715c3baa4a5116d9459993d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.292581 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.306691 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-qt56s" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97d690f0-ac29-46c9-ad4a-b92bd0a30772\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:06Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:05Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0cfd57bf783aa0e666bbe98fa7fc1eb2cda69cbd538b72fbb95491efd23e2037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-12-09T15:32:52Z\\\",\\\"message\\\":\\\"2025-12-09T15:32:06+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_fa910caf-430c-4555-b9a3-e3f63500ddee\\\\n2025-12-09T15:32:06+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_fa910caf-430c-4555-b9a3-e3f63500ddee to /host/opt/cni/bin/\\\\n2025-12-09T15:32:06Z [verbose] multus-daemon started\\\\n2025-12-09T15:32:06Z [verbose] Readiness Indicator file check\\\\n2025-12-09T15:32:51Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-12-09T15:32:05Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:53Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-czcbk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:05Z\\\"}}\" for pod \"openshift-multus\"/\"multus-qt56s\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.310768 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.310797 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.310805 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.310818 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.310826 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:06Z","lastTransitionTime":"2025-12-09T15:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.318506 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1473dc3d-be11-41e4-88d4-995a80be1e53\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:19Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0a183cf558c984ca981afe6cfbeda1cb173a511751d8edb975cf87e677e55c68\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://100782eeaa80dd7da85d32c80957a78582580bdf2bc44281f52b87865cffbe35\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r9nm5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:17Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-gsk62\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.328324 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-6hlsx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5f2fb540-7624-49a9-8143-aee00c80ae0e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b977aafca5e6308ac8a65dd0e8bb301e21ca09f49d53e00d59eb6e61e7a5a037\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:32:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cxzgx\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:32:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-6hlsx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.338037 4894 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"376c4e26-e0f9-42cb-850d-7ffc8b29e099\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:32:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-12-09T15:31:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5293b95acee426113021da42be9937be7bcefa80361ac30667ee8dad75286dee\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://882ae42ef0bd79cba2c436b8eeda122b33e1ed9efe5d09af3c80b3b1145c8103\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://b70791118fefd6a669a2e12536d1942d7120c7759a53e131f889242881918821\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-12-09T15:31:47Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://771f15a0047608775ea7c49c6fd52f868423765c9ba57a43647c830947bb0ee8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://771f15a0047608775ea7c49c6fd52f868423765c9ba57a43647c830947bb0ee8\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-12-09T15:31:47Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-12-09T15:31:46Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-12-09T15:31:46Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:06Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.413093 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.413119 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.413129 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.413143 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.413151 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:06Z","lastTransitionTime":"2025-12-09T15:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.515502 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.515557 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.515570 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.515585 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.515597 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:06Z","lastTransitionTime":"2025-12-09T15:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.617816 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.617868 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.617880 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.617897 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.617908 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:06Z","lastTransitionTime":"2025-12-09T15:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.720911 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.720959 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.720971 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.720991 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.721007 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:06Z","lastTransitionTime":"2025-12-09T15:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.823717 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.823752 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.823761 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.823776 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.823788 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:06Z","lastTransitionTime":"2025-12-09T15:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.927072 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.927126 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.927145 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.927170 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:06 crc kubenswrapper[4894]: I1209 15:33:06.927187 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:06Z","lastTransitionTime":"2025-12-09T15:33:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.030406 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.030489 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.030516 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.030548 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.030574 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:07Z","lastTransitionTime":"2025-12-09T15:33:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.106046 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.106047 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.106055 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:07 crc kubenswrapper[4894]: E1209 15:33:07.106164 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:07 crc kubenswrapper[4894]: E1209 15:33:07.106807 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:07 crc kubenswrapper[4894]: E1209 15:33:07.107125 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.132728 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.132764 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.132775 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.132790 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.132801 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:07Z","lastTransitionTime":"2025-12-09T15:33:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.234875 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.235262 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.235429 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.235599 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.235848 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:07Z","lastTransitionTime":"2025-12-09T15:33:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.338511 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.338569 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.338585 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.338609 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.338626 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:07Z","lastTransitionTime":"2025-12-09T15:33:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.441295 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.441335 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.441343 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.441356 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.441365 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:07Z","lastTransitionTime":"2025-12-09T15:33:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.543214 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.543259 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.543268 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.543282 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.543290 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:07Z","lastTransitionTime":"2025-12-09T15:33:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.645451 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.645496 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.645507 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.645524 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.645535 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:07Z","lastTransitionTime":"2025-12-09T15:33:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.748035 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.748076 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.748086 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.748109 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.748119 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:07Z","lastTransitionTime":"2025-12-09T15:33:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.849771 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.849815 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.849827 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.849842 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.849854 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:07Z","lastTransitionTime":"2025-12-09T15:33:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.952023 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.952105 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.952168 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.952189 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:07 crc kubenswrapper[4894]: I1209 15:33:07.952200 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:07Z","lastTransitionTime":"2025-12-09T15:33:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.054377 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.054408 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.054418 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.054432 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.054443 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:08Z","lastTransitionTime":"2025-12-09T15:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.106017 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:08 crc kubenswrapper[4894]: E1209 15:33:08.106144 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.157530 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.157588 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.157599 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.157617 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.157630 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:08Z","lastTransitionTime":"2025-12-09T15:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.260389 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.260421 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.260431 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.260447 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.260459 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:08Z","lastTransitionTime":"2025-12-09T15:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.363195 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.363228 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.363237 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.363251 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.363261 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:08Z","lastTransitionTime":"2025-12-09T15:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.465989 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.466151 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.466185 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.466213 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.466234 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:08Z","lastTransitionTime":"2025-12-09T15:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.569352 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.569416 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.569434 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.569458 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.569476 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:08Z","lastTransitionTime":"2025-12-09T15:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.672362 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.672417 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.672432 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.672448 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.672464 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:08Z","lastTransitionTime":"2025-12-09T15:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.774776 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.774811 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.774821 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.774835 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.774843 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:08Z","lastTransitionTime":"2025-12-09T15:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.876727 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.876782 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.876793 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.876810 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.876823 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:08Z","lastTransitionTime":"2025-12-09T15:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.911112 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.911167 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.911179 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.911198 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.911213 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:08Z","lastTransitionTime":"2025-12-09T15:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:08 crc kubenswrapper[4894]: E1209 15:33:08.924398 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.929017 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.929072 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.929087 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.929109 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.929125 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:08Z","lastTransitionTime":"2025-12-09T15:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:08 crc kubenswrapper[4894]: E1209 15:33:08.943926 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.948008 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.948053 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.948073 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.948094 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.948109 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:08Z","lastTransitionTime":"2025-12-09T15:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:08 crc kubenswrapper[4894]: E1209 15:33:08.962171 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.965807 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.965843 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.965854 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.965871 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.965881 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:08Z","lastTransitionTime":"2025-12-09T15:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:08 crc kubenswrapper[4894]: E1209 15:33:08.980302 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.984444 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.984495 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.984511 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.984534 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:08 crc kubenswrapper[4894]: I1209 15:33:08.984551 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:08Z","lastTransitionTime":"2025-12-09T15:33:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:08 crc kubenswrapper[4894]: E1209 15:33:08.999611 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404556Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865356Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"lastTransitionTime\\\":\\\"2025-12-09T15:33:08Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"a485cf49-82ef-4626-a019-8dbcfdbd7ac7\\\",\\\"systemUUID\\\":\\\"90f20501-2d99-4429-ad71-02ebb56cf15c\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-12-09T15:33:08Z is after 2025-08-24T17:21:41Z" Dec 09 15:33:08 crc kubenswrapper[4894]: E1209 15:33:08.999874 4894 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.001780 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.001823 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.001836 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.001850 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.001860 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:09Z","lastTransitionTime":"2025-12-09T15:33:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.027492 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.027743 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.027792 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:09 crc kubenswrapper[4894]: E1209 15:33:09.027832 4894 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 15:33:09 crc kubenswrapper[4894]: E1209 15:33:09.027931 4894 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 15:33:09 crc kubenswrapper[4894]: E1209 15:33:09.027832 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:13.027807902 +0000 UTC m=+147.347018581 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:33:09 crc kubenswrapper[4894]: E1209 15:33:09.028038 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:13.028016208 +0000 UTC m=+147.347226917 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Dec 09 15:33:09 crc kubenswrapper[4894]: E1209 15:33:09.028070 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:13.028055129 +0000 UTC m=+147.347265838 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.104259 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.104301 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.104313 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.104330 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.104343 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:09Z","lastTransitionTime":"2025-12-09T15:33:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.105583 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.105678 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.105588 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:09 crc kubenswrapper[4894]: E1209 15:33:09.105811 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:09 crc kubenswrapper[4894]: E1209 15:33:09.105984 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:09 crc kubenswrapper[4894]: E1209 15:33:09.106035 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.129188 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.129323 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:09 crc kubenswrapper[4894]: E1209 15:33:09.129713 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 15:33:09 crc kubenswrapper[4894]: E1209 15:33:09.129768 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 15:33:09 crc kubenswrapper[4894]: E1209 15:33:09.129795 4894 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:33:09 crc kubenswrapper[4894]: E1209 15:33:09.129897 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:13.129865807 +0000 UTC m=+147.449076516 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:33:09 crc kubenswrapper[4894]: E1209 15:33:09.130412 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Dec 09 15:33:09 crc kubenswrapper[4894]: E1209 15:33:09.130453 4894 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Dec 09 15:33:09 crc kubenswrapper[4894]: E1209 15:33:09.130491 4894 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:33:09 crc kubenswrapper[4894]: E1209 15:33:09.130566 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:13.130531754 +0000 UTC m=+147.449742463 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.213181 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.213280 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.213309 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.213342 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.213363 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:09Z","lastTransitionTime":"2025-12-09T15:33:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.315934 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.316282 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.316448 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.316882 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.316996 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:09Z","lastTransitionTime":"2025-12-09T15:33:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.420033 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.420078 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.420093 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.420114 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.420128 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:09Z","lastTransitionTime":"2025-12-09T15:33:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.523793 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.523852 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.523862 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.523878 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.523887 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:09Z","lastTransitionTime":"2025-12-09T15:33:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.625864 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.625899 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.625908 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.625920 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.625931 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:09Z","lastTransitionTime":"2025-12-09T15:33:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.728412 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.728877 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.729014 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.729142 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.729275 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:09Z","lastTransitionTime":"2025-12-09T15:33:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.832925 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.832967 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.832981 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.832999 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.833011 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:09Z","lastTransitionTime":"2025-12-09T15:33:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.935294 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.935353 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.935368 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.935386 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:09 crc kubenswrapper[4894]: I1209 15:33:09.935398 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:09Z","lastTransitionTime":"2025-12-09T15:33:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.038177 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.038218 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.038229 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.038246 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.038257 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:10Z","lastTransitionTime":"2025-12-09T15:33:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.106431 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:10 crc kubenswrapper[4894]: E1209 15:33:10.106616 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.140386 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.140426 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.140436 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.140511 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.140528 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:10Z","lastTransitionTime":"2025-12-09T15:33:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.242942 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.243001 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.243011 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.243035 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.243046 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:10Z","lastTransitionTime":"2025-12-09T15:33:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.345227 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.345265 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.345277 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.345293 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.345303 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:10Z","lastTransitionTime":"2025-12-09T15:33:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.447323 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.447354 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.447362 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.447376 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.447386 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:10Z","lastTransitionTime":"2025-12-09T15:33:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.550193 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.550245 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.550257 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.550305 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.550316 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:10Z","lastTransitionTime":"2025-12-09T15:33:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.652357 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.652418 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.652434 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.652460 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.652512 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:10Z","lastTransitionTime":"2025-12-09T15:33:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.755673 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.755724 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.755734 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.755753 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.755764 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:10Z","lastTransitionTime":"2025-12-09T15:33:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.858173 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.858253 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.858267 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.858285 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.858296 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:10Z","lastTransitionTime":"2025-12-09T15:33:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.960797 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.960836 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.960845 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.960858 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:10 crc kubenswrapper[4894]: I1209 15:33:10.960867 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:10Z","lastTransitionTime":"2025-12-09T15:33:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.064012 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.064080 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.064096 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.064115 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.064126 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:11Z","lastTransitionTime":"2025-12-09T15:33:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.105579 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.105619 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.105603 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:11 crc kubenswrapper[4894]: E1209 15:33:11.105754 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:11 crc kubenswrapper[4894]: E1209 15:33:11.105842 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:11 crc kubenswrapper[4894]: E1209 15:33:11.105894 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.165828 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.165857 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.165866 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.165878 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.165888 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:11Z","lastTransitionTime":"2025-12-09T15:33:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.267801 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.267841 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.267857 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.267874 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.267886 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:11Z","lastTransitionTime":"2025-12-09T15:33:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.370240 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.370292 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.370302 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.370318 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.370326 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:11Z","lastTransitionTime":"2025-12-09T15:33:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.472633 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.472688 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.472696 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.472710 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.472719 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:11Z","lastTransitionTime":"2025-12-09T15:33:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.576163 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.576249 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.576259 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.576313 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.576325 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:11Z","lastTransitionTime":"2025-12-09T15:33:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.682869 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.683370 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.683388 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.683404 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.683416 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:11Z","lastTransitionTime":"2025-12-09T15:33:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.786873 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.786919 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.786930 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.786948 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.786960 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:11Z","lastTransitionTime":"2025-12-09T15:33:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.889780 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.889819 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.889830 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.889845 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.889856 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:11Z","lastTransitionTime":"2025-12-09T15:33:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.992781 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.992845 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.992867 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.992896 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:11 crc kubenswrapper[4894]: I1209 15:33:11.992919 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:11Z","lastTransitionTime":"2025-12-09T15:33:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.095787 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.095869 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.095894 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.095926 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.095945 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:12Z","lastTransitionTime":"2025-12-09T15:33:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.106128 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:12 crc kubenswrapper[4894]: E1209 15:33:12.106264 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.198628 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.198699 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.198710 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.198729 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.198740 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:12Z","lastTransitionTime":"2025-12-09T15:33:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.300694 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.300748 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.300762 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.300784 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.300799 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:12Z","lastTransitionTime":"2025-12-09T15:33:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.403500 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.403788 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.403863 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.403989 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.404071 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:12Z","lastTransitionTime":"2025-12-09T15:33:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.507045 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.507098 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.507108 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.507122 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.507130 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:12Z","lastTransitionTime":"2025-12-09T15:33:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.609318 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.609570 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.609685 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.609810 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.609895 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:12Z","lastTransitionTime":"2025-12-09T15:33:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.712519 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.712866 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.712947 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.713256 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.713332 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:12Z","lastTransitionTime":"2025-12-09T15:33:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.816766 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.817057 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.817166 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.817246 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.817322 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:12Z","lastTransitionTime":"2025-12-09T15:33:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.919647 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.919730 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.919743 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.919761 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:12 crc kubenswrapper[4894]: I1209 15:33:12.919772 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:12Z","lastTransitionTime":"2025-12-09T15:33:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.021934 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.022213 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.022308 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.022401 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.022495 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:13Z","lastTransitionTime":"2025-12-09T15:33:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.105980 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.106013 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:13 crc kubenswrapper[4894]: E1209 15:33:13.106397 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.106319 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:13 crc kubenswrapper[4894]: E1209 15:33:13.106611 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:13 crc kubenswrapper[4894]: E1209 15:33:13.106619 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.124388 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.124418 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.124427 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.124440 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.124450 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:13Z","lastTransitionTime":"2025-12-09T15:33:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.226629 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.226689 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.226697 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.226710 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.226718 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:13Z","lastTransitionTime":"2025-12-09T15:33:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.328294 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.328321 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.328329 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.328341 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.328349 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:13Z","lastTransitionTime":"2025-12-09T15:33:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.430678 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.430734 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.430749 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.430765 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.430776 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:13Z","lastTransitionTime":"2025-12-09T15:33:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.533421 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.533455 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.533464 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.533478 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.533487 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:13Z","lastTransitionTime":"2025-12-09T15:33:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.635877 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.635913 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.635925 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.635942 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.635952 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:13Z","lastTransitionTime":"2025-12-09T15:33:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.738163 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.738248 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.738274 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.738288 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.738295 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:13Z","lastTransitionTime":"2025-12-09T15:33:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.842024 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.842104 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.842123 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.842152 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.842199 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:13Z","lastTransitionTime":"2025-12-09T15:33:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.944994 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.945083 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.945116 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.945148 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:13 crc kubenswrapper[4894]: I1209 15:33:13.945180 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:13Z","lastTransitionTime":"2025-12-09T15:33:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.049331 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.049395 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.049417 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.049441 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.049457 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:14Z","lastTransitionTime":"2025-12-09T15:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.105372 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:14 crc kubenswrapper[4894]: E1209 15:33:14.105630 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.106803 4894 scope.go:117] "RemoveContainer" containerID="876913572e03e323189c646a2ea4359c30b652499f20ab2c876828c0fcc5ee36" Dec 09 15:33:14 crc kubenswrapper[4894]: E1209 15:33:14.106992 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-s5scr_openshift-ovn-kubernetes(97036c6f-dce3-444b-85e6-8476c16613c9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.151971 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.152279 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.152308 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.152332 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.152346 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:14Z","lastTransitionTime":"2025-12-09T15:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.255410 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.255445 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.255456 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.255471 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.255483 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:14Z","lastTransitionTime":"2025-12-09T15:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.357676 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.357711 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.357721 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.357736 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.357746 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:14Z","lastTransitionTime":"2025-12-09T15:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.461563 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.461604 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.461613 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.461627 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.461948 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:14Z","lastTransitionTime":"2025-12-09T15:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.565269 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.565324 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.565350 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.565375 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.565390 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:14Z","lastTransitionTime":"2025-12-09T15:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.667492 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.667531 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.667540 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.667558 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.667568 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:14Z","lastTransitionTime":"2025-12-09T15:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.770248 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.770294 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.770304 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.770324 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.770336 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:14Z","lastTransitionTime":"2025-12-09T15:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.873119 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.873155 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.873166 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.873181 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.873190 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:14Z","lastTransitionTime":"2025-12-09T15:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.976199 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.976235 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.976244 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.976259 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:14 crc kubenswrapper[4894]: I1209 15:33:14.976269 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:14Z","lastTransitionTime":"2025-12-09T15:33:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.078872 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.078918 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.078927 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.078945 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.078954 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:15Z","lastTransitionTime":"2025-12-09T15:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.105664 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.105712 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:15 crc kubenswrapper[4894]: E1209 15:33:15.105835 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.105668 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:15 crc kubenswrapper[4894]: E1209 15:33:15.106030 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:15 crc kubenswrapper[4894]: E1209 15:33:15.106156 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.181519 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.181562 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.181573 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.181589 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.181600 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:15Z","lastTransitionTime":"2025-12-09T15:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.284469 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.284519 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.284530 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.284550 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.284563 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:15Z","lastTransitionTime":"2025-12-09T15:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.386881 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.386933 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.386948 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.386965 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.386976 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:15Z","lastTransitionTime":"2025-12-09T15:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.488709 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.488753 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.488764 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.488780 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.488792 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:15Z","lastTransitionTime":"2025-12-09T15:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.592522 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.592588 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.592611 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.592677 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.592702 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:15Z","lastTransitionTime":"2025-12-09T15:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.694812 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.694855 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.694865 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.694880 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.694891 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:15Z","lastTransitionTime":"2025-12-09T15:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.797696 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.797745 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.797754 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.797771 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.797781 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:15Z","lastTransitionTime":"2025-12-09T15:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.900617 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.900695 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.900706 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.900722 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:15 crc kubenswrapper[4894]: I1209 15:33:15.900733 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:15Z","lastTransitionTime":"2025-12-09T15:33:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.003574 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.003686 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.003709 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.003738 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.003760 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:16Z","lastTransitionTime":"2025-12-09T15:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.105808 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:16 crc kubenswrapper[4894]: E1209 15:33:16.105915 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.106267 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.106291 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.106301 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.106314 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.106325 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:16Z","lastTransitionTime":"2025-12-09T15:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.137846 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=16.137828024 podStartE2EDuration="16.137828024s" podCreationTimestamp="2025-12-09 15:33:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:33:16.12501492 +0000 UTC m=+90.444225589" watchObservedRunningTime="2025-12-09 15:33:16.137828024 +0000 UTC m=+90.457038683" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.192607 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-54wfs" podStartSLOduration=72.192590954 podStartE2EDuration="1m12.192590954s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:33:16.191812623 +0000 UTC m=+90.511023292" watchObservedRunningTime="2025-12-09 15:33:16.192590954 +0000 UTC m=+90.511801623" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.208478 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.208509 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.208517 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.208529 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.208539 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:16Z","lastTransitionTime":"2025-12-09T15:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.261241 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-qt56s" podStartSLOduration=72.261219816 podStartE2EDuration="1m12.261219816s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:33:16.260962379 +0000 UTC m=+90.580173058" watchObservedRunningTime="2025-12-09 15:33:16.261219816 +0000 UTC m=+90.580430495" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.285422 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-gsk62" podStartSLOduration=71.285400397 podStartE2EDuration="1m11.285400397s" podCreationTimestamp="2025-12-09 15:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:33:16.273578519 +0000 UTC m=+90.592789188" watchObservedRunningTime="2025-12-09 15:33:16.285400397 +0000 UTC m=+90.604611076" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.285572 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=40.285564381 podStartE2EDuration="40.285564381s" podCreationTimestamp="2025-12-09 15:32:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:33:16.285338035 +0000 UTC m=+90.604548704" watchObservedRunningTime="2025-12-09 15:33:16.285564381 +0000 UTC m=+90.604775080" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.310968 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.311023 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.311037 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.311053 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.311063 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:16Z","lastTransitionTime":"2025-12-09T15:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.313869 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-6hlsx" podStartSLOduration=72.3138609 podStartE2EDuration="1m12.3138609s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:33:16.296680631 +0000 UTC m=+90.615891300" watchObservedRunningTime="2025-12-09 15:33:16.3138609 +0000 UTC m=+90.633071569" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.313975 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=71.313973353 podStartE2EDuration="1m11.313973353s" podCreationTimestamp="2025-12-09 15:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:33:16.313819409 +0000 UTC m=+90.633030078" watchObservedRunningTime="2025-12-09 15:33:16.313973353 +0000 UTC m=+90.633184022" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.328045 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=72.32802587 podStartE2EDuration="1m12.32802587s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:33:16.32765921 +0000 UTC m=+90.646869899" watchObservedRunningTime="2025-12-09 15:33:16.32802587 +0000 UTC m=+90.647236549" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.352059 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-mw747" podStartSLOduration=72.352042127 podStartE2EDuration="1m12.352042127s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:33:16.351194925 +0000 UTC m=+90.670405624" watchObservedRunningTime="2025-12-09 15:33:16.352042127 +0000 UTC m=+90.671252796" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.362099 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podStartSLOduration=72.362083719 podStartE2EDuration="1m12.362083719s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:33:16.361746 +0000 UTC m=+90.680956679" watchObservedRunningTime="2025-12-09 15:33:16.362083719 +0000 UTC m=+90.681294388" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.413374 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.413410 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.413420 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.413433 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.413442 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:16Z","lastTransitionTime":"2025-12-09T15:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.515219 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.515279 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.515298 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.515320 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.515338 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:16Z","lastTransitionTime":"2025-12-09T15:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.617588 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.617624 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.617659 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.617673 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.617684 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:16Z","lastTransitionTime":"2025-12-09T15:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.719941 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.720005 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.720017 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.720033 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.720045 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:16Z","lastTransitionTime":"2025-12-09T15:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.823335 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.823377 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.823389 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.823406 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.823420 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:16Z","lastTransitionTime":"2025-12-09T15:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.926147 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.926205 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.926228 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.926249 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:16 crc kubenswrapper[4894]: I1209 15:33:16.926264 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:16Z","lastTransitionTime":"2025-12-09T15:33:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.028807 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.028853 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.028868 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.028888 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.028901 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:17Z","lastTransitionTime":"2025-12-09T15:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.105698 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.105914 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.105957 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:17 crc kubenswrapper[4894]: E1209 15:33:17.106104 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:17 crc kubenswrapper[4894]: E1209 15:33:17.106272 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:17 crc kubenswrapper[4894]: E1209 15:33:17.106408 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.131106 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.131154 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.131167 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.131186 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.131199 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:17Z","lastTransitionTime":"2025-12-09T15:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.233420 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.233485 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.233503 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.233529 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.233548 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:17Z","lastTransitionTime":"2025-12-09T15:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.336763 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.336843 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.336855 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.336881 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.336895 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:17Z","lastTransitionTime":"2025-12-09T15:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.440190 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.440275 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.440288 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.440309 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.440322 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:17Z","lastTransitionTime":"2025-12-09T15:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.543186 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.543245 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.543256 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.543282 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.543296 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:17Z","lastTransitionTime":"2025-12-09T15:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.645699 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.645743 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.645751 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.645766 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.645777 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:17Z","lastTransitionTime":"2025-12-09T15:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.748487 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.748545 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.748556 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.748575 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.748607 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:17Z","lastTransitionTime":"2025-12-09T15:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.850812 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.850857 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.850870 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.850885 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.850896 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:17Z","lastTransitionTime":"2025-12-09T15:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.955850 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.955906 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.955923 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.955943 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:17 crc kubenswrapper[4894]: I1209 15:33:17.955958 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:17Z","lastTransitionTime":"2025-12-09T15:33:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.059102 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.059169 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.059186 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.059207 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.059223 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:18Z","lastTransitionTime":"2025-12-09T15:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.105996 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:18 crc kubenswrapper[4894]: E1209 15:33:18.106197 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.162390 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.162466 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.162479 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.162493 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.162503 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:18Z","lastTransitionTime":"2025-12-09T15:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.264930 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.264977 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.264988 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.265007 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.265019 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:18Z","lastTransitionTime":"2025-12-09T15:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.368607 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.368670 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.368683 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.368700 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.368714 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:18Z","lastTransitionTime":"2025-12-09T15:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.471075 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.471133 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.471150 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.471168 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.471181 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:18Z","lastTransitionTime":"2025-12-09T15:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.573821 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.573909 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.573934 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.573963 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.573989 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:18Z","lastTransitionTime":"2025-12-09T15:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.676915 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.676951 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.676960 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.676973 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.676983 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:18Z","lastTransitionTime":"2025-12-09T15:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.779736 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.779768 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.779776 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.779789 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.779799 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:18Z","lastTransitionTime":"2025-12-09T15:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.881655 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.881707 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.881716 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.881729 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.881739 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:18Z","lastTransitionTime":"2025-12-09T15:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.984430 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.984475 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.984485 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.984501 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:18 crc kubenswrapper[4894]: I1209 15:33:18.984511 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:18Z","lastTransitionTime":"2025-12-09T15:33:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.086294 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.086347 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.086358 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.086376 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.086387 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:19Z","lastTransitionTime":"2025-12-09T15:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.105990 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.106052 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:19 crc kubenswrapper[4894]: E1209 15:33:19.106209 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.106303 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:19 crc kubenswrapper[4894]: E1209 15:33:19.106387 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:19 crc kubenswrapper[4894]: E1209 15:33:19.106504 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.189913 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.190002 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.190029 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.190070 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.190104 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:19Z","lastTransitionTime":"2025-12-09T15:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.220794 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.220853 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.220875 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.220898 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.220916 4894 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-12-09T15:33:19Z","lastTransitionTime":"2025-12-09T15:33:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.287762 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-8mhr9"] Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.288156 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8mhr9" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.290559 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.290959 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.291434 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.291939 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.340474 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/60d233fa-a5a4-4a85-b77d-363eaeb5f619-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8mhr9\" (UID: \"60d233fa-a5a4-4a85-b77d-363eaeb5f619\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8mhr9" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.340618 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/60d233fa-a5a4-4a85-b77d-363eaeb5f619-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8mhr9\" (UID: \"60d233fa-a5a4-4a85-b77d-363eaeb5f619\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8mhr9" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.340716 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60d233fa-a5a4-4a85-b77d-363eaeb5f619-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8mhr9\" (UID: \"60d233fa-a5a4-4a85-b77d-363eaeb5f619\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8mhr9" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.340839 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/60d233fa-a5a4-4a85-b77d-363eaeb5f619-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8mhr9\" (UID: \"60d233fa-a5a4-4a85-b77d-363eaeb5f619\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8mhr9" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.340908 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/60d233fa-a5a4-4a85-b77d-363eaeb5f619-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8mhr9\" (UID: \"60d233fa-a5a4-4a85-b77d-363eaeb5f619\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8mhr9" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.441610 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/60d233fa-a5a4-4a85-b77d-363eaeb5f619-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8mhr9\" (UID: \"60d233fa-a5a4-4a85-b77d-363eaeb5f619\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8mhr9" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.441780 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/60d233fa-a5a4-4a85-b77d-363eaeb5f619-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8mhr9\" (UID: \"60d233fa-a5a4-4a85-b77d-363eaeb5f619\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8mhr9" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.441831 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60d233fa-a5a4-4a85-b77d-363eaeb5f619-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8mhr9\" (UID: \"60d233fa-a5a4-4a85-b77d-363eaeb5f619\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8mhr9" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.441952 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/60d233fa-a5a4-4a85-b77d-363eaeb5f619-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8mhr9\" (UID: \"60d233fa-a5a4-4a85-b77d-363eaeb5f619\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8mhr9" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.441961 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/60d233fa-a5a4-4a85-b77d-363eaeb5f619-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-8mhr9\" (UID: \"60d233fa-a5a4-4a85-b77d-363eaeb5f619\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8mhr9" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.442021 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/60d233fa-a5a4-4a85-b77d-363eaeb5f619-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8mhr9\" (UID: \"60d233fa-a5a4-4a85-b77d-363eaeb5f619\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8mhr9" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.442147 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/60d233fa-a5a4-4a85-b77d-363eaeb5f619-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-8mhr9\" (UID: \"60d233fa-a5a4-4a85-b77d-363eaeb5f619\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8mhr9" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.442788 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/60d233fa-a5a4-4a85-b77d-363eaeb5f619-service-ca\") pod \"cluster-version-operator-5c965bbfc6-8mhr9\" (UID: \"60d233fa-a5a4-4a85-b77d-363eaeb5f619\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8mhr9" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.453246 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60d233fa-a5a4-4a85-b77d-363eaeb5f619-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-8mhr9\" (UID: \"60d233fa-a5a4-4a85-b77d-363eaeb5f619\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8mhr9" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.469577 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/60d233fa-a5a4-4a85-b77d-363eaeb5f619-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-8mhr9\" (UID: \"60d233fa-a5a4-4a85-b77d-363eaeb5f619\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8mhr9" Dec 09 15:33:19 crc kubenswrapper[4894]: I1209 15:33:19.613898 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8mhr9" Dec 09 15:33:19 crc kubenswrapper[4894]: W1209 15:33:19.636159 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60d233fa_a5a4_4a85_b77d_363eaeb5f619.slice/crio-23d0d92d85eb22b3fca381c4c73fa9333303a32be9fc29dfe3843811c6c1fa29 WatchSource:0}: Error finding container 23d0d92d85eb22b3fca381c4c73fa9333303a32be9fc29dfe3843811c6c1fa29: Status 404 returned error can't find the container with id 23d0d92d85eb22b3fca381c4c73fa9333303a32be9fc29dfe3843811c6c1fa29 Dec 09 15:33:20 crc kubenswrapper[4894]: I1209 15:33:20.106490 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:20 crc kubenswrapper[4894]: E1209 15:33:20.106693 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:20 crc kubenswrapper[4894]: I1209 15:33:20.625965 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8mhr9" event={"ID":"60d233fa-a5a4-4a85-b77d-363eaeb5f619","Type":"ContainerStarted","Data":"396c3bf4bed12414712ca501d8673cc9bea9c69e1424f3adc6806b549bcaad5a"} Dec 09 15:33:20 crc kubenswrapper[4894]: I1209 15:33:20.626020 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8mhr9" event={"ID":"60d233fa-a5a4-4a85-b77d-363eaeb5f619","Type":"ContainerStarted","Data":"23d0d92d85eb22b3fca381c4c73fa9333303a32be9fc29dfe3843811c6c1fa29"} Dec 09 15:33:20 crc kubenswrapper[4894]: I1209 15:33:20.640363 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-8mhr9" podStartSLOduration=76.640345318 podStartE2EDuration="1m16.640345318s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:33:20.640252306 +0000 UTC m=+94.959462975" watchObservedRunningTime="2025-12-09 15:33:20.640345318 +0000 UTC m=+94.959555997" Dec 09 15:33:21 crc kubenswrapper[4894]: I1209 15:33:21.106364 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:21 crc kubenswrapper[4894]: I1209 15:33:21.106399 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:21 crc kubenswrapper[4894]: I1209 15:33:21.106365 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:21 crc kubenswrapper[4894]: E1209 15:33:21.106504 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:21 crc kubenswrapper[4894]: E1209 15:33:21.106668 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:21 crc kubenswrapper[4894]: E1209 15:33:21.106762 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:22 crc kubenswrapper[4894]: I1209 15:33:22.105682 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:22 crc kubenswrapper[4894]: E1209 15:33:22.105855 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:22 crc kubenswrapper[4894]: I1209 15:33:22.878094 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs\") pod \"network-metrics-daemon-qxfr8\" (UID: \"948b28c5-e35c-4e6c-b31c-c0d5b205389b\") " pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:22 crc kubenswrapper[4894]: E1209 15:33:22.878299 4894 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 15:33:22 crc kubenswrapper[4894]: E1209 15:33:22.878546 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs podName:948b28c5-e35c-4e6c-b31c-c0d5b205389b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:26.878526273 +0000 UTC m=+161.197736952 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs") pod "network-metrics-daemon-qxfr8" (UID: "948b28c5-e35c-4e6c-b31c-c0d5b205389b") : object "openshift-multus"/"metrics-daemon-secret" not registered Dec 09 15:33:23 crc kubenswrapper[4894]: I1209 15:33:23.105544 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:23 crc kubenswrapper[4894]: I1209 15:33:23.105631 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:23 crc kubenswrapper[4894]: E1209 15:33:23.105709 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:23 crc kubenswrapper[4894]: I1209 15:33:23.105775 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:23 crc kubenswrapper[4894]: E1209 15:33:23.105887 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:23 crc kubenswrapper[4894]: E1209 15:33:23.105986 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:24 crc kubenswrapper[4894]: I1209 15:33:24.105429 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:24 crc kubenswrapper[4894]: E1209 15:33:24.105676 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:25 crc kubenswrapper[4894]: I1209 15:33:25.106495 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:25 crc kubenswrapper[4894]: I1209 15:33:25.106495 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:25 crc kubenswrapper[4894]: I1209 15:33:25.106663 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:25 crc kubenswrapper[4894]: E1209 15:33:25.106843 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:25 crc kubenswrapper[4894]: E1209 15:33:25.106946 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:25 crc kubenswrapper[4894]: E1209 15:33:25.107068 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:26 crc kubenswrapper[4894]: I1209 15:33:26.106284 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:26 crc kubenswrapper[4894]: E1209 15:33:26.107197 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:27 crc kubenswrapper[4894]: I1209 15:33:27.106231 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:27 crc kubenswrapper[4894]: I1209 15:33:27.106619 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:27 crc kubenswrapper[4894]: I1209 15:33:27.107107 4894 scope.go:117] "RemoveContainer" containerID="876913572e03e323189c646a2ea4359c30b652499f20ab2c876828c0fcc5ee36" Dec 09 15:33:27 crc kubenswrapper[4894]: I1209 15:33:27.106751 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:27 crc kubenswrapper[4894]: E1209 15:33:27.106799 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:27 crc kubenswrapper[4894]: E1209 15:33:27.107359 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-s5scr_openshift-ovn-kubernetes(97036c6f-dce3-444b-85e6-8476c16613c9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" Dec 09 15:33:27 crc kubenswrapper[4894]: E1209 15:33:27.107591 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:27 crc kubenswrapper[4894]: E1209 15:33:27.107458 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:28 crc kubenswrapper[4894]: I1209 15:33:28.106363 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:28 crc kubenswrapper[4894]: E1209 15:33:28.106541 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:29 crc kubenswrapper[4894]: I1209 15:33:29.105384 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:29 crc kubenswrapper[4894]: I1209 15:33:29.105587 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:29 crc kubenswrapper[4894]: E1209 15:33:29.105576 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:29 crc kubenswrapper[4894]: I1209 15:33:29.105616 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:29 crc kubenswrapper[4894]: E1209 15:33:29.105781 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:29 crc kubenswrapper[4894]: E1209 15:33:29.106103 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:30 crc kubenswrapper[4894]: I1209 15:33:30.106025 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:30 crc kubenswrapper[4894]: E1209 15:33:30.106362 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:31 crc kubenswrapper[4894]: I1209 15:33:31.105993 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:31 crc kubenswrapper[4894]: I1209 15:33:31.106128 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:31 crc kubenswrapper[4894]: I1209 15:33:31.105970 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:31 crc kubenswrapper[4894]: E1209 15:33:31.106332 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:31 crc kubenswrapper[4894]: E1209 15:33:31.106424 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:31 crc kubenswrapper[4894]: E1209 15:33:31.106521 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:31 crc kubenswrapper[4894]: I1209 15:33:31.122803 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Dec 09 15:33:32 crc kubenswrapper[4894]: I1209 15:33:32.105879 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:32 crc kubenswrapper[4894]: E1209 15:33:32.106085 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:33 crc kubenswrapper[4894]: I1209 15:33:33.105891 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:33 crc kubenswrapper[4894]: I1209 15:33:33.105907 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:33 crc kubenswrapper[4894]: E1209 15:33:33.106034 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:33 crc kubenswrapper[4894]: I1209 15:33:33.106058 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:33 crc kubenswrapper[4894]: E1209 15:33:33.106141 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:33 crc kubenswrapper[4894]: E1209 15:33:33.106222 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:34 crc kubenswrapper[4894]: I1209 15:33:34.106351 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:34 crc kubenswrapper[4894]: E1209 15:33:34.106488 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:35 crc kubenswrapper[4894]: I1209 15:33:35.105435 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:35 crc kubenswrapper[4894]: I1209 15:33:35.105527 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:35 crc kubenswrapper[4894]: I1209 15:33:35.105714 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:35 crc kubenswrapper[4894]: E1209 15:33:35.105794 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:35 crc kubenswrapper[4894]: E1209 15:33:35.105886 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:35 crc kubenswrapper[4894]: E1209 15:33:35.106024 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:36 crc kubenswrapper[4894]: I1209 15:33:36.106354 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:36 crc kubenswrapper[4894]: E1209 15:33:36.107672 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:36 crc kubenswrapper[4894]: I1209 15:33:36.143289 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=5.143271725 podStartE2EDuration="5.143271725s" podCreationTimestamp="2025-12-09 15:33:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:33:36.142776112 +0000 UTC m=+110.461986801" watchObservedRunningTime="2025-12-09 15:33:36.143271725 +0000 UTC m=+110.462482394" Dec 09 15:33:37 crc kubenswrapper[4894]: I1209 15:33:37.105627 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:37 crc kubenswrapper[4894]: E1209 15:33:37.105760 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:37 crc kubenswrapper[4894]: I1209 15:33:37.105669 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:37 crc kubenswrapper[4894]: E1209 15:33:37.105829 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:37 crc kubenswrapper[4894]: I1209 15:33:37.106071 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:37 crc kubenswrapper[4894]: E1209 15:33:37.106126 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:38 crc kubenswrapper[4894]: I1209 15:33:38.106511 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:38 crc kubenswrapper[4894]: E1209 15:33:38.106834 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:39 crc kubenswrapper[4894]: I1209 15:33:39.106085 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:39 crc kubenswrapper[4894]: I1209 15:33:39.106122 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:39 crc kubenswrapper[4894]: I1209 15:33:39.106231 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:39 crc kubenswrapper[4894]: E1209 15:33:39.106853 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:39 crc kubenswrapper[4894]: E1209 15:33:39.106983 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:39 crc kubenswrapper[4894]: E1209 15:33:39.107059 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:39 crc kubenswrapper[4894]: I1209 15:33:39.107351 4894 scope.go:117] "RemoveContainer" containerID="876913572e03e323189c646a2ea4359c30b652499f20ab2c876828c0fcc5ee36" Dec 09 15:33:39 crc kubenswrapper[4894]: E1209 15:33:39.107628 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-s5scr_openshift-ovn-kubernetes(97036c6f-dce3-444b-85e6-8476c16613c9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" Dec 09 15:33:39 crc kubenswrapper[4894]: I1209 15:33:39.695866 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qt56s_97d690f0-ac29-46c9-ad4a-b92bd0a30772/kube-multus/1.log" Dec 09 15:33:39 crc kubenswrapper[4894]: I1209 15:33:39.696580 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qt56s_97d690f0-ac29-46c9-ad4a-b92bd0a30772/kube-multus/0.log" Dec 09 15:33:39 crc kubenswrapper[4894]: I1209 15:33:39.696877 4894 generic.go:334] "Generic (PLEG): container finished" podID="97d690f0-ac29-46c9-ad4a-b92bd0a30772" containerID="0cfd57bf783aa0e666bbe98fa7fc1eb2cda69cbd538b72fbb95491efd23e2037" exitCode=1 Dec 09 15:33:39 crc kubenswrapper[4894]: I1209 15:33:39.696944 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qt56s" event={"ID":"97d690f0-ac29-46c9-ad4a-b92bd0a30772","Type":"ContainerDied","Data":"0cfd57bf783aa0e666bbe98fa7fc1eb2cda69cbd538b72fbb95491efd23e2037"} Dec 09 15:33:39 crc kubenswrapper[4894]: I1209 15:33:39.697273 4894 scope.go:117] "RemoveContainer" containerID="51222fb81944721d7b83a9f841c4d4943781a01cbf3d8cc744d71fcb2e6ed51e" Dec 09 15:33:39 crc kubenswrapper[4894]: I1209 15:33:39.698325 4894 scope.go:117] "RemoveContainer" containerID="0cfd57bf783aa0e666bbe98fa7fc1eb2cda69cbd538b72fbb95491efd23e2037" Dec 09 15:33:39 crc kubenswrapper[4894]: E1209 15:33:39.701511 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-qt56s_openshift-multus(97d690f0-ac29-46c9-ad4a-b92bd0a30772)\"" pod="openshift-multus/multus-qt56s" podUID="97d690f0-ac29-46c9-ad4a-b92bd0a30772" Dec 09 15:33:40 crc kubenswrapper[4894]: I1209 15:33:40.105934 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:40 crc kubenswrapper[4894]: E1209 15:33:40.106099 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:40 crc kubenswrapper[4894]: I1209 15:33:40.702802 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qt56s_97d690f0-ac29-46c9-ad4a-b92bd0a30772/kube-multus/1.log" Dec 09 15:33:41 crc kubenswrapper[4894]: I1209 15:33:41.106428 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:41 crc kubenswrapper[4894]: I1209 15:33:41.106486 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:41 crc kubenswrapper[4894]: I1209 15:33:41.106428 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:41 crc kubenswrapper[4894]: E1209 15:33:41.107045 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:41 crc kubenswrapper[4894]: E1209 15:33:41.107139 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:41 crc kubenswrapper[4894]: E1209 15:33:41.107259 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:42 crc kubenswrapper[4894]: I1209 15:33:42.105901 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:42 crc kubenswrapper[4894]: E1209 15:33:42.106150 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:43 crc kubenswrapper[4894]: I1209 15:33:43.105840 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:43 crc kubenswrapper[4894]: E1209 15:33:43.105948 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:43 crc kubenswrapper[4894]: I1209 15:33:43.106158 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:43 crc kubenswrapper[4894]: E1209 15:33:43.106220 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:43 crc kubenswrapper[4894]: I1209 15:33:43.106339 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:43 crc kubenswrapper[4894]: E1209 15:33:43.106390 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:44 crc kubenswrapper[4894]: I1209 15:33:44.105745 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:44 crc kubenswrapper[4894]: E1209 15:33:44.105895 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:45 crc kubenswrapper[4894]: I1209 15:33:45.105791 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:45 crc kubenswrapper[4894]: I1209 15:33:45.105814 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:45 crc kubenswrapper[4894]: I1209 15:33:45.106079 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:45 crc kubenswrapper[4894]: E1209 15:33:45.106149 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:45 crc kubenswrapper[4894]: E1209 15:33:45.105949 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:45 crc kubenswrapper[4894]: E1209 15:33:45.106320 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:46 crc kubenswrapper[4894]: E1209 15:33:46.080382 4894 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Dec 09 15:33:46 crc kubenswrapper[4894]: I1209 15:33:46.105623 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:46 crc kubenswrapper[4894]: E1209 15:33:46.107355 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:46 crc kubenswrapper[4894]: E1209 15:33:46.195980 4894 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 15:33:47 crc kubenswrapper[4894]: I1209 15:33:47.105863 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:47 crc kubenswrapper[4894]: I1209 15:33:47.105981 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:47 crc kubenswrapper[4894]: E1209 15:33:47.106357 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:47 crc kubenswrapper[4894]: I1209 15:33:47.106040 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:47 crc kubenswrapper[4894]: E1209 15:33:47.106523 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:47 crc kubenswrapper[4894]: E1209 15:33:47.106743 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:48 crc kubenswrapper[4894]: I1209 15:33:48.105553 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:48 crc kubenswrapper[4894]: E1209 15:33:48.105815 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:49 crc kubenswrapper[4894]: I1209 15:33:49.105850 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:49 crc kubenswrapper[4894]: I1209 15:33:49.105880 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:49 crc kubenswrapper[4894]: E1209 15:33:49.106020 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:49 crc kubenswrapper[4894]: I1209 15:33:49.106058 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:49 crc kubenswrapper[4894]: E1209 15:33:49.106175 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:49 crc kubenswrapper[4894]: E1209 15:33:49.106258 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:50 crc kubenswrapper[4894]: I1209 15:33:50.105859 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:50 crc kubenswrapper[4894]: E1209 15:33:50.105985 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:51 crc kubenswrapper[4894]: I1209 15:33:51.106055 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:51 crc kubenswrapper[4894]: I1209 15:33:51.106124 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:51 crc kubenswrapper[4894]: I1209 15:33:51.106138 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:51 crc kubenswrapper[4894]: E1209 15:33:51.106217 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:51 crc kubenswrapper[4894]: E1209 15:33:51.106344 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:51 crc kubenswrapper[4894]: E1209 15:33:51.106450 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:51 crc kubenswrapper[4894]: E1209 15:33:51.197546 4894 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Dec 09 15:33:52 crc kubenswrapper[4894]: I1209 15:33:52.106341 4894 scope.go:117] "RemoveContainer" containerID="876913572e03e323189c646a2ea4359c30b652499f20ab2c876828c0fcc5ee36" Dec 09 15:33:52 crc kubenswrapper[4894]: I1209 15:33:52.106826 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:52 crc kubenswrapper[4894]: E1209 15:33:52.106901 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:52 crc kubenswrapper[4894]: I1209 15:33:52.748617 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s5scr_97036c6f-dce3-444b-85e6-8476c16613c9/ovnkube-controller/3.log" Dec 09 15:33:52 crc kubenswrapper[4894]: I1209 15:33:52.750980 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerStarted","Data":"f27d1750c5bae1fad612cd4ef9f97c183ef4d7f6b2f9421f3b39e6fe5871a557"} Dec 09 15:33:52 crc kubenswrapper[4894]: I1209 15:33:52.751336 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:33:52 crc kubenswrapper[4894]: I1209 15:33:52.791903 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" podStartSLOduration=108.791869524 podStartE2EDuration="1m48.791869524s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:33:52.790376605 +0000 UTC m=+127.109587284" watchObservedRunningTime="2025-12-09 15:33:52.791869524 +0000 UTC m=+127.111080213" Dec 09 15:33:52 crc kubenswrapper[4894]: I1209 15:33:52.958202 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-qxfr8"] Dec 09 15:33:52 crc kubenswrapper[4894]: I1209 15:33:52.958366 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:52 crc kubenswrapper[4894]: E1209 15:33:52.958579 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:53 crc kubenswrapper[4894]: I1209 15:33:53.106375 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:53 crc kubenswrapper[4894]: I1209 15:33:53.106510 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:53 crc kubenswrapper[4894]: E1209 15:33:53.106778 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:53 crc kubenswrapper[4894]: E1209 15:33:53.106857 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:53 crc kubenswrapper[4894]: I1209 15:33:53.106905 4894 scope.go:117] "RemoveContainer" containerID="0cfd57bf783aa0e666bbe98fa7fc1eb2cda69cbd538b72fbb95491efd23e2037" Dec 09 15:33:53 crc kubenswrapper[4894]: I1209 15:33:53.756129 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qt56s_97d690f0-ac29-46c9-ad4a-b92bd0a30772/kube-multus/1.log" Dec 09 15:33:53 crc kubenswrapper[4894]: I1209 15:33:53.756245 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qt56s" event={"ID":"97d690f0-ac29-46c9-ad4a-b92bd0a30772","Type":"ContainerStarted","Data":"1b46c71319fd2c655bf490b2b8ffaf6b6a6559171f0dc8a1d30985db9e3312de"} Dec 09 15:33:54 crc kubenswrapper[4894]: I1209 15:33:54.106411 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:54 crc kubenswrapper[4894]: E1209 15:33:54.106596 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:55 crc kubenswrapper[4894]: I1209 15:33:55.105596 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:55 crc kubenswrapper[4894]: I1209 15:33:55.105675 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:55 crc kubenswrapper[4894]: E1209 15:33:55.106121 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Dec 09 15:33:55 crc kubenswrapper[4894]: E1209 15:33:55.106286 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Dec 09 15:33:55 crc kubenswrapper[4894]: I1209 15:33:55.105804 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:55 crc kubenswrapper[4894]: E1209 15:33:55.106807 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-qxfr8" podUID="948b28c5-e35c-4e6c-b31c-c0d5b205389b" Dec 09 15:33:56 crc kubenswrapper[4894]: I1209 15:33:56.106393 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:56 crc kubenswrapper[4894]: E1209 15:33:56.107290 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Dec 09 15:33:57 crc kubenswrapper[4894]: I1209 15:33:57.105992 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:33:57 crc kubenswrapper[4894]: I1209 15:33:57.106034 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:33:57 crc kubenswrapper[4894]: I1209 15:33:57.106129 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:33:57 crc kubenswrapper[4894]: I1209 15:33:57.108585 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 09 15:33:57 crc kubenswrapper[4894]: I1209 15:33:57.109221 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 09 15:33:57 crc kubenswrapper[4894]: I1209 15:33:57.109366 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 09 15:33:57 crc kubenswrapper[4894]: I1209 15:33:57.108915 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 09 15:33:58 crc kubenswrapper[4894]: I1209 15:33:58.105852 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:33:58 crc kubenswrapper[4894]: I1209 15:33:58.110114 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 09 15:33:58 crc kubenswrapper[4894]: I1209 15:33:58.110130 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.898514 4894 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.937268 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826"] Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.937818 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc"] Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.938398 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.938961 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.939320 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-rvzr9"] Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.939819 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.940519 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fmq8p"] Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.940867 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fmq8p" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.941610 4894 reflector.go:561] object-"openshift-oauth-apiserver"/"audit-1": failed to list *v1.ConfigMap: configmaps "audit-1" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.941680 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"audit-1\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"audit-1\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.941726 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-bcfbr"] Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.942388 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.944363 4894 reflector.go:561] object-"openshift-oauth-apiserver"/"encryption-config-1": failed to list *v1.Secret: secrets "encryption-config-1" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.944396 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"encryption-config-1\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"encryption-config-1\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.944854 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2thb4"] Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.945671 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2thb4" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.946121 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qjlt6"] Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.946483 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.947247 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-8bhhd"] Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.947742 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8bhhd" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.948617 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-w6q9w"] Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.948961 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-w6q9w" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.950117 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-crhvg"] Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.950485 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.962461 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-x4lf2"] Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.962684 4894 reflector.go:561] object-"openshift-route-controller-manager"/"client-ca": failed to list *v1.ConfigMap: configmaps "client-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.962736 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"client-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"client-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.962750 4894 reflector.go:561] object-"openshift-oauth-apiserver"/"etcd-client": failed to list *v1.Secret: secrets "etcd-client" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.962792 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"etcd-client\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"etcd-client\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.964506 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-x4lf2" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.965142 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22kg8"] Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.979230 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22kg8" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.980566 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-vtbc8"] Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.980589 4894 reflector.go:561] object-"openshift-oauth-apiserver"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.980628 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.980634 4894 reflector.go:561] object-"openshift-route-controller-manager"/"config": failed to list *v1.ConfigMap: configmaps "config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.980687 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.980756 4894 reflector.go:561] object-"openshift-route-controller-manager"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.980774 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.980912 4894 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-tls": failed to list *v1.Secret: secrets "machine-api-operator-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.980954 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-api-operator-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.980963 4894 reflector.go:561] object-"openshift-oauth-apiserver"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.980989 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.980916 4894 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7": failed to list *v1.Secret: secrets "machine-api-operator-dockercfg-mfbb7" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.981018 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-dockercfg-mfbb7\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"machine-api-operator-dockercfg-mfbb7\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.981050 4894 reflector.go:561] object-"openshift-machine-api"/"kube-rbac-proxy": failed to list *v1.ConfigMap: configmaps "kube-rbac-proxy" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.981070 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"kube-rbac-proxy\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-rbac-proxy\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.981098 4894 reflector.go:561] object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq": failed to list *v1.Secret: secrets "oauth-apiserver-sa-dockercfg-6r2bq" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.981116 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"oauth-apiserver-sa-dockercfg-6r2bq\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"oauth-apiserver-sa-dockercfg-6r2bq\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.981134 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-vtbc8" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.981169 4894 reflector.go:561] object-"openshift-machine-api"/"machine-api-operator-images": failed to list *v1.ConfigMap: configmaps "machine-api-operator-images" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.981183 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"machine-api-operator-images\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"machine-api-operator-images\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.981218 4894 reflector.go:561] object-"openshift-oauth-apiserver"/"trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.981239 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.981400 4894 reflector.go:561] object-"openshift-apiserver"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.981421 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.981663 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.981707 4894 reflector.go:561] object-"openshift-cluster-samples-operator"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-samples-operator": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.981729 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-samples-operator\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-samples-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.981739 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.981849 4894 reflector.go:561] object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2": failed to list *v1.Secret: secrets "route-controller-manager-sa-dockercfg-h2zr2" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.981870 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"route-controller-manager-sa-dockercfg-h2zr2\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"route-controller-manager-sa-dockercfg-h2zr2\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.981882 4894 reflector.go:561] object-"openshift-apiserver"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.981926 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.981884 4894 reflector.go:561] object-"openshift-route-controller-manager"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.981950 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.981963 4894 reflector.go:561] object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w": failed to list *v1.Secret: secrets "cluster-samples-operator-dockercfg-xpp9w" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-cluster-samples-operator": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.981981 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.981983 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-samples-operator\"/\"cluster-samples-operator-dockercfg-xpp9w\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cluster-samples-operator-dockercfg-xpp9w\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-cluster-samples-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.982027 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.982092 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.982125 4894 reflector.go:561] object-"openshift-route-controller-manager"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-route-controller-manager": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.982147 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.982146 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-route-controller-manager\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-route-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.982135 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.982262 4894 reflector.go:561] object-"openshift-apiserver"/"etcd-client": failed to list *v1.Secret: secrets "etcd-client" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.982279 4894 reflector.go:561] object-"openshift-machine-api"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.982282 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"etcd-client\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"etcd-client\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.982296 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.982323 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.982569 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.982584 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.982697 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.982885 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.982980 4894 reflector.go:561] object-"openshift-oauth-apiserver"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.982980 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-j5jl6"] Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.982998 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.983135 4894 reflector.go:561] object-"openshift-machine-api"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-machine-api": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.983153 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-machine-api\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-machine-api\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.983314 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.983431 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.983451 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.983509 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-j5jl6" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.983516 4894 reflector.go:561] object-"openshift-cluster-samples-operator"/"samples-operator-tls": failed to list *v1.Secret: secrets "samples-operator-tls" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-cluster-samples-operator": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.983883 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-samples-operator\"/\"samples-operator-tls\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"samples-operator-tls\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-cluster-samples-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.983553 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.984021 4894 reflector.go:561] object-"openshift-apiserver"/"etcd-serving-ca": failed to list *v1.ConfigMap: configmaps "etcd-serving-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.984044 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"etcd-serving-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"etcd-serving-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.984083 4894 reflector.go:561] object-"openshift-oauth-apiserver"/"etcd-serving-ca": failed to list *v1.ConfigMap: configmaps "etcd-serving-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-oauth-apiserver": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.984096 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-oauth-apiserver\"/\"etcd-serving-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"etcd-serving-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-oauth-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.984133 4894 reflector.go:561] object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff": failed to list *v1.Secret: secrets "openshift-apiserver-sa-dockercfg-djjff" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.984145 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"openshift-apiserver-sa-dockercfg-djjff\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"openshift-apiserver-sa-dockercfg-djjff\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.984193 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.983573 4894 reflector.go:561] object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-cluster-samples-operator": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.984315 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-cluster-samples-operator\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-cluster-samples-operator\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.983590 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.983626 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.984417 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.983693 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.983700 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.984543 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.983759 4894 reflector.go:561] object-"openshift-apiserver"/"image-import-ca": failed to list *v1.ConfigMap: configmaps "image-import-ca" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.984576 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"image-import-ca\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"image-import-ca\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.983800 4894 reflector.go:561] object-"openshift-apiserver"/"trusted-ca-bundle": failed to list *v1.ConfigMap: configmaps "trusted-ca-bundle" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.984600 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"trusted-ca-bundle\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"trusted-ca-bundle\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.983847 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 09 15:33:59 crc kubenswrapper[4894]: W1209 15:33:59.984675 4894 reflector.go:561] object-"openshift-apiserver"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-apiserver": no relationship found between node 'crc' and this object Dec 09 15:33:59 crc kubenswrapper[4894]: E1209 15:33:59.984698 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-apiserver\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.984838 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.984987 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.985087 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.985106 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.985129 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.985097 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.984333 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.985243 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.985322 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.985365 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.984295 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.985451 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.985538 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.985548 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.987691 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k8jmv"] Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.988350 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.988905 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ks8v2"] Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.989681 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ks8v2" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.990449 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6n52l"] Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.992625 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6n52l" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.993652 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nd9jv\" (UniqueName: \"kubernetes.io/projected/c83b2004-58fd-4b02-b989-9566ee5fd283-kube-api-access-nd9jv\") pod \"dns-operator-744455d44c-x4lf2\" (UID: \"c83b2004-58fd-4b02-b989-9566ee5fd283\") " pod="openshift-dns-operator/dns-operator-744455d44c-x4lf2" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.993714 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66f45f17-addc-4737-a47c-1ab34a289659-config\") pod \"machine-api-operator-5694c8668f-rvzr9\" (UID: \"66f45f17-addc-4737-a47c-1ab34a289659\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.993750 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-oauth-serving-cert\") pod \"console-f9d7485db-crhvg\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.993788 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qd294\" (UniqueName: \"kubernetes.io/projected/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-kube-api-access-qd294\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.993857 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x646t\" (UniqueName: \"kubernetes.io/projected/0d089cc1-fe1c-4ad4-a707-27c0e809649a-kube-api-access-x646t\") pod \"cluster-samples-operator-665b6dd947-fmq8p\" (UID: \"0d089cc1-fe1c-4ad4-a707-27c0e809649a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fmq8p" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.993897 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-serving-cert\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.993934 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-node-pullsecrets\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.993973 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqfdg\" (UniqueName: \"kubernetes.io/projected/00d36531-cc42-49a2-a4c9-28f9a5ddd044-kube-api-access-qqfdg\") pod \"controller-manager-879f6c89f-qjlt6\" (UID: \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.994015 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/abf8acce-d250-4d2d-895f-3d59afee8b73-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2thb4\" (UID: \"abf8acce-d250-4d2d-895f-3d59afee8b73\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2thb4" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.994057 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d548d70-c4af-4333-8cea-f6bef5722ce5-console-serving-cert\") pod \"console-f9d7485db-crhvg\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.994117 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zz7dd\" (UniqueName: \"kubernetes.io/projected/b3748054-88ec-46d8-af2e-15565a8b772a-kube-api-access-zz7dd\") pod \"authentication-operator-69f744f599-w6q9w\" (UID: \"b3748054-88ec-46d8-af2e-15565a8b772a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w6q9w" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.994154 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzmm4\" (UniqueName: \"kubernetes.io/projected/9d548d70-c4af-4333-8cea-f6bef5722ce5-kube-api-access-kzmm4\") pod \"console-f9d7485db-crhvg\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.994196 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00d36531-cc42-49a2-a4c9-28f9a5ddd044-config\") pod \"controller-manager-879f6c89f-qjlt6\" (UID: \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.994242 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qj9dp\" (UniqueName: \"kubernetes.io/projected/66f45f17-addc-4737-a47c-1ab34a289659-kube-api-access-qj9dp\") pod \"machine-api-operator-5694c8668f-rvzr9\" (UID: \"66f45f17-addc-4737-a47c-1ab34a289659\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.994299 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3748054-88ec-46d8-af2e-15565a8b772a-serving-cert\") pod \"authentication-operator-69f744f599-w6q9w\" (UID: \"b3748054-88ec-46d8-af2e-15565a8b772a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w6q9w" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.994350 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d548d70-c4af-4333-8cea-f6bef5722ce5-console-oauth-config\") pod \"console-f9d7485db-crhvg\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.994410 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-trusted-ca-bundle\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.994462 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/901f250c-9e0d-4901-ac4d-0c8c336e67b8-machine-approver-tls\") pod \"machine-approver-56656f9798-8bhhd\" (UID: \"901f250c-9e0d-4901-ac4d-0c8c336e67b8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8bhhd" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.994530 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/66f45f17-addc-4737-a47c-1ab34a289659-images\") pod \"machine-api-operator-5694c8668f-rvzr9\" (UID: \"66f45f17-addc-4737-a47c-1ab34a289659\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.994571 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b3748054-88ec-46d8-af2e-15565a8b772a-service-ca-bundle\") pod \"authentication-operator-69f744f599-w6q9w\" (UID: \"b3748054-88ec-46d8-af2e-15565a8b772a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w6q9w" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.994605 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-config\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.994666 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/00d36531-cc42-49a2-a4c9-28f9a5ddd044-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qjlt6\" (UID: \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.994707 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/901f250c-9e0d-4901-ac4d-0c8c336e67b8-config\") pod \"machine-approver-56656f9798-8bhhd\" (UID: \"901f250c-9e0d-4901-ac4d-0c8c336e67b8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8bhhd" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.994732 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9sjj\" (UniqueName: \"kubernetes.io/projected/171c863a-36f5-48c0-be81-752684537f0b-kube-api-access-h9sjj\") pod \"route-controller-manager-6576b87f9c-56826\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.994764 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/00d36531-cc42-49a2-a4c9-28f9a5ddd044-client-ca\") pod \"controller-manager-879f6c89f-qjlt6\" (UID: \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.994802 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-client-ca\") pod \"route-controller-manager-6576b87f9c-56826\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.994837 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0d089cc1-fe1c-4ad4-a707-27c0e809649a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fmq8p\" (UID: \"0d089cc1-fe1c-4ad4-a707-27c0e809649a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fmq8p" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.994878 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-audit\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.994912 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-service-ca\") pod \"console-f9d7485db-crhvg\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.994959 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00d36531-cc42-49a2-a4c9-28f9a5ddd044-serving-cert\") pod \"controller-manager-879f6c89f-qjlt6\" (UID: \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995002 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/171c863a-36f5-48c0-be81-752684537f0b-serving-cert\") pod \"route-controller-manager-6576b87f9c-56826\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995047 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-msz24\" (UniqueName: \"kubernetes.io/projected/abf8acce-d250-4d2d-895f-3d59afee8b73-kube-api-access-msz24\") pod \"openshift-apiserver-operator-796bbdcf4f-2thb4\" (UID: \"abf8acce-d250-4d2d-895f-3d59afee8b73\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2thb4" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995086 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-console-config\") pod \"console-f9d7485db-crhvg\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995121 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-encryption-config\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995161 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-config\") pod \"route-controller-manager-6576b87f9c-56826\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995194 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abf8acce-d250-4d2d-895f-3d59afee8b73-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2thb4\" (UID: \"abf8acce-d250-4d2d-895f-3d59afee8b73\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2thb4" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995234 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-serving-cert\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995273 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c83b2004-58fd-4b02-b989-9566ee5fd283-metrics-tls\") pod \"dns-operator-744455d44c-x4lf2\" (UID: \"c83b2004-58fd-4b02-b989-9566ee5fd283\") " pod="openshift-dns-operator/dns-operator-744455d44c-x4lf2" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995315 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f3b629b0-15d9-4e1d-908c-b21f07444427-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995355 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f3b629b0-15d9-4e1d-908c-b21f07444427-audit-policies\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995434 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-etcd-client\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995518 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f3b629b0-15d9-4e1d-908c-b21f07444427-audit-dir\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995564 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-audit-dir\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995592 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f3b629b0-15d9-4e1d-908c-b21f07444427-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995615 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-etcd-client\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995678 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hg269\" (UniqueName: \"kubernetes.io/projected/901f250c-9e0d-4901-ac4d-0c8c336e67b8-kube-api-access-hg269\") pod \"machine-approver-56656f9798-8bhhd\" (UID: \"901f250c-9e0d-4901-ac4d-0c8c336e67b8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8bhhd" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995378 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8n5bj"] Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995818 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3748054-88ec-46d8-af2e-15565a8b772a-config\") pod \"authentication-operator-69f744f599-w6q9w\" (UID: \"b3748054-88ec-46d8-af2e-15565a8b772a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w6q9w" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995841 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b3748054-88ec-46d8-af2e-15565a8b772a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-w6q9w\" (UID: \"b3748054-88ec-46d8-af2e-15565a8b772a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w6q9w" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995861 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq8cc\" (UniqueName: \"kubernetes.io/projected/f3b629b0-15d9-4e1d-908c-b21f07444427-kube-api-access-xq8cc\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995879 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-etcd-serving-ca\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995895 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/901f250c-9e0d-4901-ac4d-0c8c336e67b8-auth-proxy-config\") pod \"machine-approver-56656f9798-8bhhd\" (UID: \"901f250c-9e0d-4901-ac4d-0c8c336e67b8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8bhhd" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995910 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-trusted-ca-bundle\") pod \"console-f9d7485db-crhvg\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995931 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/66f45f17-addc-4737-a47c-1ab34a289659-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-rvzr9\" (UID: \"66f45f17-addc-4737-a47c-1ab34a289659\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995946 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-encryption-config\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.995962 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-image-import-ca\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.996236 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8n5bj" Dec 09 15:33:59 crc kubenswrapper[4894]: I1209 15:33:59.997326 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.000706 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-g79bw"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.001313 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.001429 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.001943 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rjzql"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.002381 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rjzql" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.003956 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.004123 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dsfnf"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.004677 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.005177 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.006773 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4cv2"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.007809 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4cv2" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.007965 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd8c7"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.008556 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd8c7" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.008580 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.009263 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.009471 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.009498 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.009693 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.010026 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.010577 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.010621 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.010771 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.010857 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.011107 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.011288 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.011456 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.026544 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-d7q9r"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.027615 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-d7q9r" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.028095 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.028315 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.028474 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.028633 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.028982 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.029531 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.030147 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.053903 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.054127 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.054908 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.055033 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.056202 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.056381 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.059313 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.059448 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5px2s"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.060795 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.064839 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.065151 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sfrz4"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.065519 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5px2s" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.065618 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sfrz4" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.065871 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.066753 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.066962 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.066977 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.077049 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z94kw"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.077717 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z94kw" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.080138 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ktwf9"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.080819 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ktwf9" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.082085 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7nlp9"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.082895 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-7nlp9" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.083451 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m8gtv"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.084022 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m8gtv" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.086422 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-sd4tb"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.087311 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-sd4tb" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.090058 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-lcnnn"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.090753 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qzp8"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.091121 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qzp8" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.091378 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lcnnn" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.093295 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fmq8p"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.095237 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2thb4"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.097070 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/62c287c6-7ccb-4d05-8e30-c83ea71bb9d6-etcd-client\") pod \"etcd-operator-b45778765-dsfnf\" (UID: \"62c287c6-7ccb-4d05-8e30-c83ea71bb9d6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.097119 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/66f45f17-addc-4737-a47c-1ab34a289659-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-rvzr9\" (UID: \"66f45f17-addc-4737-a47c-1ab34a289659\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.097147 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-encryption-config\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.097172 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/62c287c6-7ccb-4d05-8e30-c83ea71bb9d6-etcd-service-ca\") pod \"etcd-operator-b45778765-dsfnf\" (UID: \"62c287c6-7ccb-4d05-8e30-c83ea71bb9d6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.097195 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-image-import-ca\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.097216 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q25gn\" (UniqueName: \"kubernetes.io/projected/215f4af3-778b-4657-a9c2-dca7fd769055-kube-api-access-q25gn\") pod \"olm-operator-6b444d44fb-z94kw\" (UID: \"215f4af3-778b-4657-a9c2-dca7fd769055\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z94kw" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.097327 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66f45f17-addc-4737-a47c-1ab34a289659-config\") pod \"machine-api-operator-5694c8668f-rvzr9\" (UID: \"66f45f17-addc-4737-a47c-1ab34a289659\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.097357 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-oauth-serving-cert\") pod \"console-f9d7485db-crhvg\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.097379 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qd294\" (UniqueName: \"kubernetes.io/projected/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-kube-api-access-qd294\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.097421 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nd9jv\" (UniqueName: \"kubernetes.io/projected/c83b2004-58fd-4b02-b989-9566ee5fd283-kube-api-access-nd9jv\") pod \"dns-operator-744455d44c-x4lf2\" (UID: \"c83b2004-58fd-4b02-b989-9566ee5fd283\") " pod="openshift-dns-operator/dns-operator-744455d44c-x4lf2" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.097446 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x646t\" (UniqueName: \"kubernetes.io/projected/0d089cc1-fe1c-4ad4-a707-27c0e809649a-kube-api-access-x646t\") pod \"cluster-samples-operator-665b6dd947-fmq8p\" (UID: \"0d089cc1-fe1c-4ad4-a707-27c0e809649a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fmq8p" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.097478 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-node-pullsecrets\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.097499 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqfdg\" (UniqueName: \"kubernetes.io/projected/00d36531-cc42-49a2-a4c9-28f9a5ddd044-kube-api-access-qqfdg\") pod \"controller-manager-879f6c89f-qjlt6\" (UID: \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.097628 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/abf8acce-d250-4d2d-895f-3d59afee8b73-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2thb4\" (UID: \"abf8acce-d250-4d2d-895f-3d59afee8b73\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2thb4" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.097740 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-serving-cert\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.097765 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d548d70-c4af-4333-8cea-f6bef5722ce5-console-serving-cert\") pod \"console-f9d7485db-crhvg\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098039 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/62c287c6-7ccb-4d05-8e30-c83ea71bb9d6-etcd-ca\") pod \"etcd-operator-b45778765-dsfnf\" (UID: \"62c287c6-7ccb-4d05-8e30-c83ea71bb9d6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098085 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cwqsk\" (UniqueName: \"kubernetes.io/projected/62c287c6-7ccb-4d05-8e30-c83ea71bb9d6-kube-api-access-cwqsk\") pod \"etcd-operator-b45778765-dsfnf\" (UID: \"62c287c6-7ccb-4d05-8e30-c83ea71bb9d6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098110 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zz7dd\" (UniqueName: \"kubernetes.io/projected/b3748054-88ec-46d8-af2e-15565a8b772a-kube-api-access-zz7dd\") pod \"authentication-operator-69f744f599-w6q9w\" (UID: \"b3748054-88ec-46d8-af2e-15565a8b772a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w6q9w" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098132 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzmm4\" (UniqueName: \"kubernetes.io/projected/9d548d70-c4af-4333-8cea-f6bef5722ce5-kube-api-access-kzmm4\") pod \"console-f9d7485db-crhvg\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098154 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00d36531-cc42-49a2-a4c9-28f9a5ddd044-config\") pod \"controller-manager-879f6c89f-qjlt6\" (UID: \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098239 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qj9dp\" (UniqueName: \"kubernetes.io/projected/66f45f17-addc-4737-a47c-1ab34a289659-kube-api-access-qj9dp\") pod \"machine-api-operator-5694c8668f-rvzr9\" (UID: \"66f45f17-addc-4737-a47c-1ab34a289659\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098321 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3748054-88ec-46d8-af2e-15565a8b772a-serving-cert\") pod \"authentication-operator-69f744f599-w6q9w\" (UID: \"b3748054-88ec-46d8-af2e-15565a8b772a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w6q9w" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098349 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d548d70-c4af-4333-8cea-f6bef5722ce5-console-oauth-config\") pod \"console-f9d7485db-crhvg\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098369 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-trusted-ca-bundle\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098394 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/901f250c-9e0d-4901-ac4d-0c8c336e67b8-machine-approver-tls\") pod \"machine-approver-56656f9798-8bhhd\" (UID: \"901f250c-9e0d-4901-ac4d-0c8c336e67b8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8bhhd" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098414 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/66f45f17-addc-4737-a47c-1ab34a289659-images\") pod \"machine-api-operator-5694c8668f-rvzr9\" (UID: \"66f45f17-addc-4737-a47c-1ab34a289659\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098444 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/448379d9-ce78-4910-8651-541f26b80f4d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-22kg8\" (UID: \"448379d9-ce78-4910-8651-541f26b80f4d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22kg8" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098466 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62c287c6-7ccb-4d05-8e30-c83ea71bb9d6-serving-cert\") pod \"etcd-operator-b45778765-dsfnf\" (UID: \"62c287c6-7ccb-4d05-8e30-c83ea71bb9d6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098491 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b3748054-88ec-46d8-af2e-15565a8b772a-service-ca-bundle\") pod \"authentication-operator-69f744f599-w6q9w\" (UID: \"b3748054-88ec-46d8-af2e-15565a8b772a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w6q9w" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098511 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-config\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098535 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/00d36531-cc42-49a2-a4c9-28f9a5ddd044-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qjlt6\" (UID: \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098558 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/901f250c-9e0d-4901-ac4d-0c8c336e67b8-config\") pod \"machine-approver-56656f9798-8bhhd\" (UID: \"901f250c-9e0d-4901-ac4d-0c8c336e67b8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8bhhd" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098584 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9sjj\" (UniqueName: \"kubernetes.io/projected/171c863a-36f5-48c0-be81-752684537f0b-kube-api-access-h9sjj\") pod \"route-controller-manager-6576b87f9c-56826\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098610 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/448379d9-ce78-4910-8651-541f26b80f4d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-22kg8\" (UID: \"448379d9-ce78-4910-8651-541f26b80f4d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22kg8" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098654 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-client-ca\") pod \"route-controller-manager-6576b87f9c-56826\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098680 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0d089cc1-fe1c-4ad4-a707-27c0e809649a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fmq8p\" (UID: \"0d089cc1-fe1c-4ad4-a707-27c0e809649a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fmq8p" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098699 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/00d36531-cc42-49a2-a4c9-28f9a5ddd044-client-ca\") pod \"controller-manager-879f6c89f-qjlt6\" (UID: \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098720 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-audit\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098742 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-service-ca\") pod \"console-f9d7485db-crhvg\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098765 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zvnbf\" (UniqueName: \"kubernetes.io/projected/448379d9-ce78-4910-8651-541f26b80f4d-kube-api-access-zvnbf\") pod \"cluster-image-registry-operator-dc59b4c8b-22kg8\" (UID: \"448379d9-ce78-4910-8651-541f26b80f4d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22kg8" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098792 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/215f4af3-778b-4657-a9c2-dca7fd769055-profile-collector-cert\") pod \"olm-operator-6b444d44fb-z94kw\" (UID: \"215f4af3-778b-4657-a9c2-dca7fd769055\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z94kw" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098817 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/171c863a-36f5-48c0-be81-752684537f0b-serving-cert\") pod \"route-controller-manager-6576b87f9c-56826\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098843 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00d36531-cc42-49a2-a4c9-28f9a5ddd044-serving-cert\") pod \"controller-manager-879f6c89f-qjlt6\" (UID: \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098865 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-console-config\") pod \"console-f9d7485db-crhvg\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098887 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-encryption-config\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098911 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-msz24\" (UniqueName: \"kubernetes.io/projected/abf8acce-d250-4d2d-895f-3d59afee8b73-kube-api-access-msz24\") pod \"openshift-apiserver-operator-796bbdcf4f-2thb4\" (UID: \"abf8acce-d250-4d2d-895f-3d59afee8b73\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2thb4" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098937 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-config\") pod \"route-controller-manager-6576b87f9c-56826\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098959 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abf8acce-d250-4d2d-895f-3d59afee8b73-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2thb4\" (UID: \"abf8acce-d250-4d2d-895f-3d59afee8b73\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2thb4" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.098981 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-serving-cert\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099003 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/448379d9-ce78-4910-8651-541f26b80f4d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-22kg8\" (UID: \"448379d9-ce78-4910-8651-541f26b80f4d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22kg8" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099025 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss5fc\" (UniqueName: \"kubernetes.io/projected/caed7380-e514-4f40-ae98-0be4a2fe61c3-kube-api-access-ss5fc\") pod \"downloads-7954f5f757-vtbc8\" (UID: \"caed7380-e514-4f40-ae98-0be4a2fe61c3\") " pod="openshift-console/downloads-7954f5f757-vtbc8" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099055 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c83b2004-58fd-4b02-b989-9566ee5fd283-metrics-tls\") pod \"dns-operator-744455d44c-x4lf2\" (UID: \"c83b2004-58fd-4b02-b989-9566ee5fd283\") " pod="openshift-dns-operator/dns-operator-744455d44c-x4lf2" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099079 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f3b629b0-15d9-4e1d-908c-b21f07444427-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099101 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/215f4af3-778b-4657-a9c2-dca7fd769055-srv-cert\") pod \"olm-operator-6b444d44fb-z94kw\" (UID: \"215f4af3-778b-4657-a9c2-dca7fd769055\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z94kw" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099124 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f3b629b0-15d9-4e1d-908c-b21f07444427-audit-policies\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099148 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-etcd-client\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099145 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-x4lf2"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099230 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-audit-dir\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099264 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62c287c6-7ccb-4d05-8e30-c83ea71bb9d6-config\") pod \"etcd-operator-b45778765-dsfnf\" (UID: \"62c287c6-7ccb-4d05-8e30-c83ea71bb9d6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099290 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f3b629b0-15d9-4e1d-908c-b21f07444427-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099314 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f3b629b0-15d9-4e1d-908c-b21f07444427-audit-dir\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099337 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hg269\" (UniqueName: \"kubernetes.io/projected/901f250c-9e0d-4901-ac4d-0c8c336e67b8-kube-api-access-hg269\") pod \"machine-approver-56656f9798-8bhhd\" (UID: \"901f250c-9e0d-4901-ac4d-0c8c336e67b8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8bhhd" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099373 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-etcd-client\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099398 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3748054-88ec-46d8-af2e-15565a8b772a-config\") pod \"authentication-operator-69f744f599-w6q9w\" (UID: \"b3748054-88ec-46d8-af2e-15565a8b772a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w6q9w" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099423 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b3748054-88ec-46d8-af2e-15565a8b772a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-w6q9w\" (UID: \"b3748054-88ec-46d8-af2e-15565a8b772a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w6q9w" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099447 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq8cc\" (UniqueName: \"kubernetes.io/projected/f3b629b0-15d9-4e1d-908c-b21f07444427-kube-api-access-xq8cc\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099473 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-etcd-serving-ca\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099560 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/901f250c-9e0d-4901-ac4d-0c8c336e67b8-auth-proxy-config\") pod \"machine-approver-56656f9798-8bhhd\" (UID: \"901f250c-9e0d-4901-ac4d-0c8c336e67b8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8bhhd" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099706 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-trusted-ca-bundle\") pod \"console-f9d7485db-crhvg\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099764 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099883 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-audit\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.099949 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/901f250c-9e0d-4901-ac4d-0c8c336e67b8-config\") pod \"machine-approver-56656f9798-8bhhd\" (UID: \"901f250c-9e0d-4901-ac4d-0c8c336e67b8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8bhhd" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.101210 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b3748054-88ec-46d8-af2e-15565a8b772a-service-ca-bundle\") pod \"authentication-operator-69f744f599-w6q9w\" (UID: \"b3748054-88ec-46d8-af2e-15565a8b772a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w6q9w" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.101257 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-service-ca\") pod \"console-f9d7485db-crhvg\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.101353 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/00d36531-cc42-49a2-a4c9-28f9a5ddd044-client-ca\") pod \"controller-manager-879f6c89f-qjlt6\" (UID: \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.101507 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f3b629b0-15d9-4e1d-908c-b21f07444427-audit-dir\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.101548 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/abf8acce-d250-4d2d-895f-3d59afee8b73-config\") pod \"openshift-apiserver-operator-796bbdcf4f-2thb4\" (UID: \"abf8acce-d250-4d2d-895f-3d59afee8b73\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2thb4" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.101761 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/00d36531-cc42-49a2-a4c9-28f9a5ddd044-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-qjlt6\" (UID: \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.101784 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-trusted-ca-bundle\") pod \"console-f9d7485db-crhvg\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.101881 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-console-config\") pod \"console-f9d7485db-crhvg\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.102514 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/901f250c-9e0d-4901-ac4d-0c8c336e67b8-auth-proxy-config\") pod \"machine-approver-56656f9798-8bhhd\" (UID: \"901f250c-9e0d-4901-ac4d-0c8c336e67b8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8bhhd" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.102538 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-config\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.102802 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-84l9b"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.102862 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b3748054-88ec-46d8-af2e-15565a8b772a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-w6q9w\" (UID: \"b3748054-88ec-46d8-af2e-15565a8b772a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w6q9w" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.103554 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-84l9b" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.103817 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-crhvg"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.104997 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-oauth-serving-cert\") pod \"console-f9d7485db-crhvg\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.105582 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b3748054-88ec-46d8-af2e-15565a8b772a-config\") pod \"authentication-operator-69f744f599-w6q9w\" (UID: \"b3748054-88ec-46d8-af2e-15565a8b772a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w6q9w" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.107170 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-node-pullsecrets\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.107302 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-audit-dir\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.107591 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/abf8acce-d250-4d2d-895f-3d59afee8b73-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-2thb4\" (UID: \"abf8acce-d250-4d2d-895f-3d59afee8b73\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2thb4" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.108909 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00d36531-cc42-49a2-a4c9-28f9a5ddd044-config\") pod \"controller-manager-879f6c89f-qjlt6\" (UID: \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.111101 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00d36531-cc42-49a2-a4c9-28f9a5ddd044-serving-cert\") pod \"controller-manager-879f6c89f-qjlt6\" (UID: \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.111662 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d548d70-c4af-4333-8cea-f6bef5722ce5-console-serving-cert\") pod \"console-f9d7485db-crhvg\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.112074 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-encryption-config\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.116346 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/c83b2004-58fd-4b02-b989-9566ee5fd283-metrics-tls\") pod \"dns-operator-744455d44c-x4lf2\" (UID: \"c83b2004-58fd-4b02-b989-9566ee5fd283\") " pod="openshift-dns-operator/dns-operator-744455d44c-x4lf2" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.117075 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/901f250c-9e0d-4901-ac4d-0c8c336e67b8-machine-approver-tls\") pod \"machine-approver-56656f9798-8bhhd\" (UID: \"901f250c-9e0d-4901-ac4d-0c8c336e67b8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8bhhd" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.122457 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.131487 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b3748054-88ec-46d8-af2e-15565a8b772a-serving-cert\") pod \"authentication-operator-69f744f599-w6q9w\" (UID: \"b3748054-88ec-46d8-af2e-15565a8b772a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w6q9w" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.135805 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.138326 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d548d70-c4af-4333-8cea-f6bef5722ce5-console-oauth-config\") pod \"console-f9d7485db-crhvg\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.143934 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.144842 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qq9k2"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.145619 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.146063 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gr92k"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.146616 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.146739 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.147038 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-j5jl6"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.147067 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-g79bw"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.147089 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5px2s"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.147102 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ktwf9"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.147148 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ks8v2"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.147159 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-bcfbr"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.147168 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-d7q9r"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.147178 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22kg8"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.147236 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gr92k" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.148733 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z94kw"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.150976 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-vtbc8"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.152245 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qjlt6"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.157097 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6n52l"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.158559 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-84l9b"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.160370 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-95rf5"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.160854 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.162395 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-95rf5" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.163257 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-rvzr9"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.164610 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-lcnnn"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.167106 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.169579 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-d5smt"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.171373 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-d5smt" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.173139 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4cv2"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.175790 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k8jmv"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.177956 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sfrz4"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.181146 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.181940 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd8c7"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.184996 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rjzql"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.188921 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7nlp9"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.190626 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-w6q9w"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.192606 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8n5bj"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.194016 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qq9k2"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.195141 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qzp8"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.196767 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dsfnf"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.198327 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m8gtv"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.200530 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.200549 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/448379d9-ce78-4910-8651-541f26b80f4d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-22kg8\" (UID: \"448379d9-ce78-4910-8651-541f26b80f4d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22kg8" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.200683 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ss5fc\" (UniqueName: \"kubernetes.io/projected/caed7380-e514-4f40-ae98-0be4a2fe61c3-kube-api-access-ss5fc\") pod \"downloads-7954f5f757-vtbc8\" (UID: \"caed7380-e514-4f40-ae98-0be4a2fe61c3\") " pod="openshift-console/downloads-7954f5f757-vtbc8" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.200714 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/215f4af3-778b-4657-a9c2-dca7fd769055-srv-cert\") pod \"olm-operator-6b444d44fb-z94kw\" (UID: \"215f4af3-778b-4657-a9c2-dca7fd769055\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z94kw" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.200773 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62c287c6-7ccb-4d05-8e30-c83ea71bb9d6-config\") pod \"etcd-operator-b45778765-dsfnf\" (UID: \"62c287c6-7ccb-4d05-8e30-c83ea71bb9d6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.200852 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/62c287c6-7ccb-4d05-8e30-c83ea71bb9d6-etcd-service-ca\") pod \"etcd-operator-b45778765-dsfnf\" (UID: \"62c287c6-7ccb-4d05-8e30-c83ea71bb9d6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.200879 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/62c287c6-7ccb-4d05-8e30-c83ea71bb9d6-etcd-client\") pod \"etcd-operator-b45778765-dsfnf\" (UID: \"62c287c6-7ccb-4d05-8e30-c83ea71bb9d6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.200885 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-d5smt"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.200905 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q25gn\" (UniqueName: \"kubernetes.io/projected/215f4af3-778b-4657-a9c2-dca7fd769055-kube-api-access-q25gn\") pod \"olm-operator-6b444d44fb-z94kw\" (UID: \"215f4af3-778b-4657-a9c2-dca7fd769055\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z94kw" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.201562 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/62c287c6-7ccb-4d05-8e30-c83ea71bb9d6-etcd-ca\") pod \"etcd-operator-b45778765-dsfnf\" (UID: \"62c287c6-7ccb-4d05-8e30-c83ea71bb9d6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.201599 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cwqsk\" (UniqueName: \"kubernetes.io/projected/62c287c6-7ccb-4d05-8e30-c83ea71bb9d6-kube-api-access-cwqsk\") pod \"etcd-operator-b45778765-dsfnf\" (UID: \"62c287c6-7ccb-4d05-8e30-c83ea71bb9d6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.201677 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/448379d9-ce78-4910-8651-541f26b80f4d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-22kg8\" (UID: \"448379d9-ce78-4910-8651-541f26b80f4d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22kg8" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.201708 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62c287c6-7ccb-4d05-8e30-c83ea71bb9d6-serving-cert\") pod \"etcd-operator-b45778765-dsfnf\" (UID: \"62c287c6-7ccb-4d05-8e30-c83ea71bb9d6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.201745 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/448379d9-ce78-4910-8651-541f26b80f4d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-22kg8\" (UID: \"448379d9-ce78-4910-8651-541f26b80f4d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22kg8" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.201781 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zvnbf\" (UniqueName: \"kubernetes.io/projected/448379d9-ce78-4910-8651-541f26b80f4d-kube-api-access-zvnbf\") pod \"cluster-image-registry-operator-dc59b4c8b-22kg8\" (UID: \"448379d9-ce78-4910-8651-541f26b80f4d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22kg8" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.201792 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-l6cb5"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.201810 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/215f4af3-778b-4657-a9c2-dca7fd769055-profile-collector-cert\") pod \"olm-operator-6b444d44fb-z94kw\" (UID: \"215f4af3-778b-4657-a9c2-dca7fd769055\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z94kw" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.202906 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-cjsx7"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.203309 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-cjsx7" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.203621 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.203968 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/448379d9-ce78-4910-8651-541f26b80f4d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-22kg8\" (UID: \"448379d9-ce78-4910-8651-541f26b80f4d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22kg8" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.204079 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.205060 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/448379d9-ce78-4910-8651-541f26b80f4d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-22kg8\" (UID: \"448379d9-ce78-4910-8651-541f26b80f4d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22kg8" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.205269 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-cjsx7"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.206295 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-l6cb5"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.207299 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gr92k"] Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.220389 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.240545 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.275069 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.287096 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.301133 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.321156 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.340267 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.361628 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.381545 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.400517 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.421543 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.441279 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.447062 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/62c287c6-7ccb-4d05-8e30-c83ea71bb9d6-serving-cert\") pod \"etcd-operator-b45778765-dsfnf\" (UID: \"62c287c6-7ccb-4d05-8e30-c83ea71bb9d6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.461087 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.464094 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/62c287c6-7ccb-4d05-8e30-c83ea71bb9d6-etcd-client\") pod \"etcd-operator-b45778765-dsfnf\" (UID: \"62c287c6-7ccb-4d05-8e30-c83ea71bb9d6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.481156 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.492215 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/62c287c6-7ccb-4d05-8e30-c83ea71bb9d6-config\") pod \"etcd-operator-b45778765-dsfnf\" (UID: \"62c287c6-7ccb-4d05-8e30-c83ea71bb9d6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.502216 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.512725 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/62c287c6-7ccb-4d05-8e30-c83ea71bb9d6-etcd-ca\") pod \"etcd-operator-b45778765-dsfnf\" (UID: \"62c287c6-7ccb-4d05-8e30-c83ea71bb9d6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.521828 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.532141 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/62c287c6-7ccb-4d05-8e30-c83ea71bb9d6-etcd-service-ca\") pod \"etcd-operator-b45778765-dsfnf\" (UID: \"62c287c6-7ccb-4d05-8e30-c83ea71bb9d6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.540904 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.560696 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.581431 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.602275 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.622568 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.641584 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.662176 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.681883 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.702086 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.720839 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.740538 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.763401 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.781725 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.801332 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.842292 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.860064 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.882309 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.900999 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.922018 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.941289 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.962100 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 09 15:34:00 crc kubenswrapper[4894]: I1209 15:34:00.981043 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.001901 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.022505 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.041746 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.045982 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/215f4af3-778b-4657-a9c2-dca7fd769055-profile-collector-cert\") pod \"olm-operator-6b444d44fb-z94kw\" (UID: \"215f4af3-778b-4657-a9c2-dca7fd769055\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z94kw" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.061261 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.074968 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/215f4af3-778b-4657-a9c2-dca7fd769055-srv-cert\") pod \"olm-operator-6b444d44fb-z94kw\" (UID: \"215f4af3-778b-4657-a9c2-dca7fd769055\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z94kw" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.099034 4894 request.go:700] Waited for 1.0179736s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-operator-lifecycle-manager/secrets?fieldSelector=metadata.name%3Dcatalog-operator-serving-cert&limit=500&resourceVersion=0 Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.100272 4894 secret.go:188] Couldn't get secret openshift-route-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.100299 4894 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/etcd-serving-ca: failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.100369 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/171c863a-36f5-48c0-be81-752684537f0b-serving-cert podName:171c863a-36f5-48c0-be81-752684537f0b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:01.60034268 +0000 UTC m=+135.919553389 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/171c863a-36f5-48c0-be81-752684537f0b-serving-cert") pod "route-controller-manager-6576b87f9c-56826" (UID: "171c863a-36f5-48c0-be81-752684537f0b") : failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.100409 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f3b629b0-15d9-4e1d-908c-b21f07444427-etcd-serving-ca podName:f3b629b0-15d9-4e1d-908c-b21f07444427 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:01.600394601 +0000 UTC m=+135.919605310 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-serving-ca" (UniqueName: "kubernetes.io/configmap/f3b629b0-15d9-4e1d-908c-b21f07444427-etcd-serving-ca") pod "apiserver-7bbb656c7d-2qdjc" (UID: "f3b629b0-15d9-4e1d-908c-b21f07444427") : failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.100414 4894 secret.go:188] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.100414 4894 configmap.go:193] Couldn't get configMap openshift-machine-api/machine-api-operator-images: failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.100437 4894 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.100502 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0d089cc1-fe1c-4ad4-a707-27c0e809649a-samples-operator-tls podName:0d089cc1-fe1c-4ad4-a707-27c0e809649a nodeName:}" failed. No retries permitted until 2025-12-09 15:34:01.600480913 +0000 UTC m=+135.919691592 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/0d089cc1-fe1c-4ad4-a707-27c0e809649a-samples-operator-tls") pod "cluster-samples-operator-665b6dd947-fmq8p" (UID: "0d089cc1-fe1c-4ad4-a707-27c0e809649a") : failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.100376 4894 secret.go:188] Couldn't get secret openshift-oauth-apiserver/encryption-config-1: failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.100526 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/66f45f17-addc-4737-a47c-1ab34a289659-images podName:66f45f17-addc-4737-a47c-1ab34a289659 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:01.600512414 +0000 UTC m=+135.919723093 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "images" (UniqueName: "kubernetes.io/configmap/66f45f17-addc-4737-a47c-1ab34a289659-images") pod "machine-api-operator-5694c8668f-rvzr9" (UID: "66f45f17-addc-4737-a47c-1ab34a289659") : failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.100539 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-client-ca podName:171c863a-36f5-48c0-be81-752684537f0b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:01.600532215 +0000 UTC m=+135.919742894 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-client-ca") pod "route-controller-manager-6576b87f9c-56826" (UID: "171c863a-36f5-48c0-be81-752684537f0b") : failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.100567 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-encryption-config podName:f3b629b0-15d9-4e1d-908c-b21f07444427 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:01.600546175 +0000 UTC m=+135.919756874 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "encryption-config" (UniqueName: "kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-encryption-config") pod "apiserver-7bbb656c7d-2qdjc" (UID: "f3b629b0-15d9-4e1d-908c-b21f07444427") : failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.101949 4894 configmap.go:193] Couldn't get configMap openshift-apiserver/etcd-serving-ca: failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.102005 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-etcd-serving-ca podName:39bada6f-31fa-46b6-bd31-6f43cc0f6aa4 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:01.601990394 +0000 UTC m=+135.921201083 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-serving-ca" (UniqueName: "kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-etcd-serving-ca") pod "apiserver-76f77b778f-bcfbr" (UID: "39bada6f-31fa-46b6-bd31-6f43cc0f6aa4") : failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.102003 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.102001 4894 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.102116 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-config podName:171c863a-36f5-48c0-be81-752684537f0b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:01.602093106 +0000 UTC m=+135.921303815 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-config") pod "route-controller-manager-6576b87f9c-56826" (UID: "171c863a-36f5-48c0-be81-752684537f0b") : failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.102205 4894 secret.go:188] Couldn't get secret openshift-machine-api/machine-api-operator-tls: failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.102244 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/66f45f17-addc-4737-a47c-1ab34a289659-machine-api-operator-tls podName:66f45f17-addc-4737-a47c-1ab34a289659 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:01.60223292 +0000 UTC m=+135.921443599 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "machine-api-operator-tls" (UniqueName: "kubernetes.io/secret/66f45f17-addc-4737-a47c-1ab34a289659-machine-api-operator-tls") pod "machine-api-operator-5694c8668f-rvzr9" (UID: "66f45f17-addc-4737-a47c-1ab34a289659") : failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.103445 4894 secret.go:188] Couldn't get secret openshift-apiserver/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.103511 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-serving-cert podName:39bada6f-31fa-46b6-bd31-6f43cc0f6aa4 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:01.603494754 +0000 UTC m=+135.922705463 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-serving-cert") pod "apiserver-76f77b778f-bcfbr" (UID: "39bada6f-31fa-46b6-bd31-6f43cc0f6aa4") : failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.103538 4894 secret.go:188] Couldn't get secret openshift-oauth-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.103577 4894 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/audit-1: failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.103667 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-etcd-client podName:f3b629b0-15d9-4e1d-908c-b21f07444427 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:01.603612647 +0000 UTC m=+135.922823356 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-etcd-client") pod "apiserver-7bbb656c7d-2qdjc" (UID: "f3b629b0-15d9-4e1d-908c-b21f07444427") : failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.103711 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f3b629b0-15d9-4e1d-908c-b21f07444427-audit-policies podName:f3b629b0-15d9-4e1d-908c-b21f07444427 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:01.603687579 +0000 UTC m=+135.922898348 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "audit-policies" (UniqueName: "kubernetes.io/configmap/f3b629b0-15d9-4e1d-908c-b21f07444427-audit-policies") pod "apiserver-7bbb656c7d-2qdjc" (UID: "f3b629b0-15d9-4e1d-908c-b21f07444427") : failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.103863 4894 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.104061 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f3b629b0-15d9-4e1d-908c-b21f07444427-trusted-ca-bundle podName:f3b629b0-15d9-4e1d-908c-b21f07444427 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:01.603927885 +0000 UTC m=+135.923138594 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/f3b629b0-15d9-4e1d-908c-b21f07444427-trusted-ca-bundle") pod "apiserver-7bbb656c7d-2qdjc" (UID: "f3b629b0-15d9-4e1d-908c-b21f07444427") : failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.105345 4894 configmap.go:193] Couldn't get configMap openshift-apiserver/image-import-ca: failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.105421 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-image-import-ca podName:39bada6f-31fa-46b6-bd31-6f43cc0f6aa4 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:01.605404225 +0000 UTC m=+135.924614934 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "image-import-ca" (UniqueName: "kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-image-import-ca") pod "apiserver-76f77b778f-bcfbr" (UID: "39bada6f-31fa-46b6-bd31-6f43cc0f6aa4") : failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.105355 4894 configmap.go:193] Couldn't get configMap openshift-machine-api/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.105472 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/66f45f17-addc-4737-a47c-1ab34a289659-config podName:66f45f17-addc-4737-a47c-1ab34a289659 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:01.605460376 +0000 UTC m=+135.924671145 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/66f45f17-addc-4737-a47c-1ab34a289659-config") pod "machine-api-operator-5694c8668f-rvzr9" (UID: "66f45f17-addc-4737-a47c-1ab34a289659") : failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.108417 4894 secret.go:188] Couldn't get secret openshift-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.108512 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-etcd-client podName:39bada6f-31fa-46b6-bd31-6f43cc0f6aa4 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:01.608486927 +0000 UTC m=+135.927697656 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-etcd-client") pod "apiserver-76f77b778f-bcfbr" (UID: "39bada6f-31fa-46b6-bd31-6f43cc0f6aa4") : failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.108514 4894 configmap.go:193] Couldn't get configMap openshift-apiserver/trusted-ca-bundle: failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.108570 4894 secret.go:188] Couldn't get secret openshift-oauth-apiserver/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.108600 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-trusted-ca-bundle podName:39bada6f-31fa-46b6-bd31-6f43cc0f6aa4 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:01.608577191 +0000 UTC m=+135.927787920 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "trusted-ca-bundle" (UniqueName: "kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-trusted-ca-bundle") pod "apiserver-76f77b778f-bcfbr" (UID: "39bada6f-31fa-46b6-bd31-6f43cc0f6aa4") : failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: E1209 15:34:01.108684 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-serving-cert podName:f3b629b0-15d9-4e1d-908c-b21f07444427 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:01.608668333 +0000 UTC m=+135.927879092 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-serving-cert") pod "apiserver-7bbb656c7d-2qdjc" (UID: "f3b629b0-15d9-4e1d-908c-b21f07444427") : failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.121674 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.140530 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.161179 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.181907 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.200450 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.220750 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.241142 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.262537 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.281292 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.301758 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.321213 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.341938 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.361907 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.380897 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.439379 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hg269\" (UniqueName: \"kubernetes.io/projected/901f250c-9e0d-4901-ac4d-0c8c336e67b8-kube-api-access-hg269\") pod \"machine-approver-56656f9798-8bhhd\" (UID: \"901f250c-9e0d-4901-ac4d-0c8c336e67b8\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8bhhd" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.458598 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-msz24\" (UniqueName: \"kubernetes.io/projected/abf8acce-d250-4d2d-895f-3d59afee8b73-kube-api-access-msz24\") pod \"openshift-apiserver-operator-796bbdcf4f-2thb4\" (UID: \"abf8acce-d250-4d2d-895f-3d59afee8b73\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2thb4" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.476251 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqfdg\" (UniqueName: \"kubernetes.io/projected/00d36531-cc42-49a2-a4c9-28f9a5ddd044-kube-api-access-qqfdg\") pod \"controller-manager-879f6c89f-qjlt6\" (UID: \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\") " pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.502270 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.521935 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.554456 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zz7dd\" (UniqueName: \"kubernetes.io/projected/b3748054-88ec-46d8-af2e-15565a8b772a-kube-api-access-zz7dd\") pod \"authentication-operator-69f744f599-w6q9w\" (UID: \"b3748054-88ec-46d8-af2e-15565a8b772a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-w6q9w" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.575143 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nd9jv\" (UniqueName: \"kubernetes.io/projected/c83b2004-58fd-4b02-b989-9566ee5fd283-kube-api-access-nd9jv\") pod \"dns-operator-744455d44c-x4lf2\" (UID: \"c83b2004-58fd-4b02-b989-9566ee5fd283\") " pod="openshift-dns-operator/dns-operator-744455d44c-x4lf2" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.620836 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.621125 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/66f45f17-addc-4737-a47c-1ab34a289659-images\") pod \"machine-api-operator-5694c8668f-rvzr9\" (UID: \"66f45f17-addc-4737-a47c-1ab34a289659\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.621177 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-client-ca\") pod \"route-controller-manager-6576b87f9c-56826\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.621201 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0d089cc1-fe1c-4ad4-a707-27c0e809649a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fmq8p\" (UID: \"0d089cc1-fe1c-4ad4-a707-27c0e809649a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fmq8p" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.621231 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/171c863a-36f5-48c0-be81-752684537f0b-serving-cert\") pod \"route-controller-manager-6576b87f9c-56826\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.621250 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-encryption-config\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.621266 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-serving-cert\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.621289 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-config\") pod \"route-controller-manager-6576b87f9c-56826\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.621306 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f3b629b0-15d9-4e1d-908c-b21f07444427-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.621322 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f3b629b0-15d9-4e1d-908c-b21f07444427-audit-policies\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.621339 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-etcd-client\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.621364 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f3b629b0-15d9-4e1d-908c-b21f07444427-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.621381 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-etcd-client\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.621409 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-etcd-serving-ca\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.621427 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/66f45f17-addc-4737-a47c-1ab34a289659-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-rvzr9\" (UID: \"66f45f17-addc-4737-a47c-1ab34a289659\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.621446 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-image-import-ca\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.621473 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66f45f17-addc-4737-a47c-1ab34a289659-config\") pod \"machine-api-operator-5694c8668f-rvzr9\" (UID: \"66f45f17-addc-4737-a47c-1ab34a289659\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.621508 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-serving-cert\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.621550 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-trusted-ca-bundle\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.638989 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2thb4" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.649308 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.657507 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8bhhd" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.666936 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-w6q9w" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.677865 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzmm4\" (UniqueName: \"kubernetes.io/projected/9d548d70-c4af-4333-8cea-f6bef5722ce5-kube-api-access-kzmm4\") pod \"console-f9d7485db-crhvg\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.678955 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.681740 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.689460 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-x4lf2" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.702053 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.720902 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.740831 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.765156 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.791519 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.795107 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8bhhd" event={"ID":"901f250c-9e0d-4901-ac4d-0c8c336e67b8","Type":"ContainerStarted","Data":"c1ed800fe9fa7b7ddcef8c9f41e5d3e54ecb44a1d6d06b8f93162ebd6eb4157f"} Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.801093 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.821610 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.842553 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.861394 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.885774 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.901340 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.903491 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qjlt6"] Dec 09 15:34:01 crc kubenswrapper[4894]: W1209 15:34:01.913889 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00d36531_cc42_49a2_a4c9_28f9a5ddd044.slice/crio-adff7418b9342665f8a003a2f969034eec7663da81d7dda35af281f14d4f5cda WatchSource:0}: Error finding container adff7418b9342665f8a003a2f969034eec7663da81d7dda35af281f14d4f5cda: Status 404 returned error can't find the container with id adff7418b9342665f8a003a2f969034eec7663da81d7dda35af281f14d4f5cda Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.923987 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.940374 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.962164 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 09 15:34:01 crc kubenswrapper[4894]: I1209 15:34:01.981095 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.001537 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.021014 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.040877 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.061011 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.096600 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/448379d9-ce78-4910-8651-541f26b80f4d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-22kg8\" (UID: \"448379d9-ce78-4910-8651-541f26b80f4d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22kg8" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.099787 4894 request.go:700] Waited for 1.898928636s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-console/serviceaccounts/default/token Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.124127 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ss5fc\" (UniqueName: \"kubernetes.io/projected/caed7380-e514-4f40-ae98-0be4a2fe61c3-kube-api-access-ss5fc\") pod \"downloads-7954f5f757-vtbc8\" (UID: \"caed7380-e514-4f40-ae98-0be4a2fe61c3\") " pod="openshift-console/downloads-7954f5f757-vtbc8" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.130235 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.137680 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-crhvg"] Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.138971 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-w6q9w"] Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.140580 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-x4lf2"] Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.142185 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2thb4"] Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.148244 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q25gn\" (UniqueName: \"kubernetes.io/projected/215f4af3-778b-4657-a9c2-dca7fd769055-kube-api-access-q25gn\") pod \"olm-operator-6b444d44fb-z94kw\" (UID: \"215f4af3-778b-4657-a9c2-dca7fd769055\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z94kw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.148473 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z94kw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.163594 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cwqsk\" (UniqueName: \"kubernetes.io/projected/62c287c6-7ccb-4d05-8e30-c83ea71bb9d6-kube-api-access-cwqsk\") pod \"etcd-operator-b45778765-dsfnf\" (UID: \"62c287c6-7ccb-4d05-8e30-c83ea71bb9d6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.176315 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zvnbf\" (UniqueName: \"kubernetes.io/projected/448379d9-ce78-4910-8651-541f26b80f4d-kube-api-access-zvnbf\") pod \"cluster-image-registry-operator-dc59b4c8b-22kg8\" (UID: \"448379d9-ce78-4910-8651-541f26b80f4d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22kg8" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.180753 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.200333 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.221514 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.240897 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.261397 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.280866 4894 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.295678 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22kg8" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.301660 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.302698 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-vtbc8" Dec 09 15:34:02 crc kubenswrapper[4894]: W1209 15:34:02.326919 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3748054_88ec_46d8_af2e_15565a8b772a.slice/crio-2ff7ec84129e06e1373743999368acc74c2f679c35104655afeaa221f82396ea WatchSource:0}: Error finding container 2ff7ec84129e06e1373743999368acc74c2f679c35104655afeaa221f82396ea: Status 404 returned error can't find the container with id 2ff7ec84129e06e1373743999368acc74c2f679c35104655afeaa221f82396ea Dec 09 15:34:02 crc kubenswrapper[4894]: W1209 15:34:02.330293 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d548d70_c4af_4333_8cea_f6bef5722ce5.slice/crio-3a9d5ae85cba8e07af05cd9ed81836acb2a42edd0fafb15888d327bd0fe8f5f5 WatchSource:0}: Error finding container 3a9d5ae85cba8e07af05cd9ed81836acb2a42edd0fafb15888d327bd0fe8f5f5: Status 404 returned error can't find the container with id 3a9d5ae85cba8e07af05cd9ed81836acb2a42edd0fafb15888d327bd0fe8f5f5 Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.341782 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.360360 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.363764 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-image-import-ca\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.379260 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.386763 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.392760 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-trusted-ca-bundle\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.401952 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.411627 4894 projected.go:288] Couldn't get configMap openshift-route-controller-manager/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.421847 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.441239 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.466882 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.481524 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.482321 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f3b629b0-15d9-4e1d-908c-b21f07444427-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.486117 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x646t\" (UniqueName: \"kubernetes.io/projected/0d089cc1-fe1c-4ad4-a707-27c0e809649a-kube-api-access-x646t\") pod \"cluster-samples-operator-665b6dd947-fmq8p\" (UID: \"0d089cc1-fe1c-4ad4-a707-27c0e809649a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fmq8p" Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.494175 4894 projected.go:288] Couldn't get configMap openshift-oauth-apiserver/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.500239 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.523044 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.541738 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.566454 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.573060 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f3b629b0-15d9-4e1d-908c-b21f07444427-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.580826 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.587106 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-etcd-client\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.593590 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22kg8"] Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.606894 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z94kw"] Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.607096 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.618867 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/66f45f17-addc-4737-a47c-1ab34a289659-images\") pod \"machine-api-operator-5694c8668f-rvzr9\" (UID: \"66f45f17-addc-4737-a47c-1ab34a289659\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.629685 4894 configmap.go:193] Couldn't get configMap openshift-machine-api/kube-rbac-proxy: failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.629746 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/66f45f17-addc-4737-a47c-1ab34a289659-config podName:66f45f17-addc-4737-a47c-1ab34a289659 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.62972901 +0000 UTC m=+137.948939679 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/66f45f17-addc-4737-a47c-1ab34a289659-config") pod "machine-api-operator-5694c8668f-rvzr9" (UID: "66f45f17-addc-4737-a47c-1ab34a289659") : failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.629769 4894 secret.go:188] Couldn't get secret openshift-oauth-apiserver/encryption-config-1: failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.629875 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-encryption-config podName:f3b629b0-15d9-4e1d-908c-b21f07444427 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.629847603 +0000 UTC m=+137.949058342 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "encryption-config" (UniqueName: "kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-encryption-config") pod "apiserver-7bbb656c7d-2qdjc" (UID: "f3b629b0-15d9-4e1d-908c-b21f07444427") : failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.629907 4894 secret.go:188] Couldn't get secret openshift-route-controller-manager/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.629938 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/171c863a-36f5-48c0-be81-752684537f0b-serving-cert podName:171c863a-36f5-48c0-be81-752684537f0b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.629930536 +0000 UTC m=+137.949141305 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/171c863a-36f5-48c0-be81-752684537f0b-serving-cert") pod "route-controller-manager-6576b87f9c-56826" (UID: "171c863a-36f5-48c0-be81-752684537f0b") : failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.629953 4894 secret.go:188] Couldn't get secret openshift-cluster-samples-operator/samples-operator-tls: failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.629976 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0d089cc1-fe1c-4ad4-a707-27c0e809649a-samples-operator-tls podName:0d089cc1-fe1c-4ad4-a707-27c0e809649a nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.629970857 +0000 UTC m=+137.949181526 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "samples-operator-tls" (UniqueName: "kubernetes.io/secret/0d089cc1-fe1c-4ad4-a707-27c0e809649a-samples-operator-tls") pod "cluster-samples-operator-665b6dd947-fmq8p" (UID: "0d089cc1-fe1c-4ad4-a707-27c0e809649a") : failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.629994 4894 secret.go:188] Couldn't get secret openshift-oauth-apiserver/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.630023 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-serving-cert podName:f3b629b0-15d9-4e1d-908c-b21f07444427 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.630016018 +0000 UTC m=+137.949226687 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-serving-cert") pod "apiserver-7bbb656c7d-2qdjc" (UID: "f3b629b0-15d9-4e1d-908c-b21f07444427") : failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.630023 4894 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/config: failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.630039 4894 secret.go:188] Couldn't get secret openshift-apiserver/serving-cert: failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.630065 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-serving-cert podName:39bada6f-31fa-46b6-bd31-6f43cc0f6aa4 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.630060549 +0000 UTC m=+137.949271218 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "serving-cert" (UniqueName: "kubernetes.io/secret/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-serving-cert") pod "apiserver-76f77b778f-bcfbr" (UID: "39bada6f-31fa-46b6-bd31-6f43cc0f6aa4") : failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.630081 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-config podName:171c863a-36f5-48c0-be81-752684537f0b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.630075659 +0000 UTC m=+137.949286328 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-config") pod "route-controller-manager-6576b87f9c-56826" (UID: "171c863a-36f5-48c0-be81-752684537f0b") : failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.630102 4894 secret.go:188] Couldn't get secret openshift-apiserver/etcd-client: failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.630105 4894 configmap.go:193] Couldn't get configMap openshift-route-controller-manager/client-ca: failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.630123 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-etcd-client podName:39bada6f-31fa-46b6-bd31-6f43cc0f6aa4 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.6301176 +0000 UTC m=+137.949328269 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etcd-client" (UniqueName: "kubernetes.io/secret/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-etcd-client") pod "apiserver-76f77b778f-bcfbr" (UID: "39bada6f-31fa-46b6-bd31-6f43cc0f6aa4") : failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.630138 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-client-ca podName:171c863a-36f5-48c0-be81-752684537f0b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.630132351 +0000 UTC m=+137.949343020 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "client-ca" (UniqueName: "kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-client-ca") pod "route-controller-manager-6576b87f9c-56826" (UID: "171c863a-36f5-48c0-be81-752684537f0b") : failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.630145 4894 secret.go:188] Couldn't get secret openshift-machine-api/machine-api-operator-tls: failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.630159 4894 configmap.go:193] Couldn't get configMap openshift-apiserver/etcd-serving-ca: failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.630166 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/66f45f17-addc-4737-a47c-1ab34a289659-machine-api-operator-tls podName:66f45f17-addc-4737-a47c-1ab34a289659 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.630161202 +0000 UTC m=+137.949371871 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "machine-api-operator-tls" (UniqueName: "kubernetes.io/secret/66f45f17-addc-4737-a47c-1ab34a289659-machine-api-operator-tls") pod "machine-api-operator-5694c8668f-rvzr9" (UID: "66f45f17-addc-4737-a47c-1ab34a289659") : failed to sync secret cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.630177 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-etcd-serving-ca podName:39bada6f-31fa-46b6-bd31-6f43cc0f6aa4 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.630172252 +0000 UTC m=+137.949382921 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etcd-serving-ca" (UniqueName: "kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-etcd-serving-ca") pod "apiserver-76f77b778f-bcfbr" (UID: "39bada6f-31fa-46b6-bd31-6f43cc0f6aa4") : failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.630186 4894 configmap.go:193] Couldn't get configMap openshift-oauth-apiserver/audit-1: failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.630205 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/f3b629b0-15d9-4e1d-908c-b21f07444427-audit-policies podName:f3b629b0-15d9-4e1d-908c-b21f07444427 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.630200783 +0000 UTC m=+137.949411452 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "audit-policies" (UniqueName: "kubernetes.io/configmap/f3b629b0-15d9-4e1d-908c-b21f07444427-audit-policies") pod "apiserver-7bbb656c7d-2qdjc" (UID: "f3b629b0-15d9-4e1d-908c-b21f07444427") : failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.640750 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/661b7c44-b4e8-4c2c-a7b8-a752564f557b-images\") pod \"machine-config-operator-74547568cd-5px2s\" (UID: \"661b7c44-b4e8-4c2c-a7b8-a752564f557b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5px2s" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.640795 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f15a4328-85fb-49db-aed5-24a76ce8cc0e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-sd8c7\" (UID: \"f15a4328-85fb-49db-aed5-24a76ce8cc0e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd8c7" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.640834 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-audit-policies\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.640859 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/661b7c44-b4e8-4c2c-a7b8-a752564f557b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5px2s\" (UID: \"661b7c44-b4e8-4c2c-a7b8-a752564f557b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5px2s" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.640883 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/661b7c44-b4e8-4c2c-a7b8-a752564f557b-proxy-tls\") pod \"machine-config-operator-74547568cd-5px2s\" (UID: \"661b7c44-b4e8-4c2c-a7b8-a752564f557b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5px2s" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.640914 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/adab756d-1eb4-4837-a93a-346a38f83df9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.640960 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8b4b2ae6-2dca-4267-b764-d6ca08e601ed-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-8n5bj\" (UID: \"8b4b2ae6-2dca-4267-b764-d6ca08e601ed\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8n5bj" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.640982 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1ec98242-7220-4547-b699-5e97bddfbca9-webhook-cert\") pod \"packageserver-d55dfcdfc-pqs8f\" (UID: \"1ec98242-7220-4547-b699-5e97bddfbca9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.641035 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.641054 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.641086 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.641112 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.641156 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.641178 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s5wxh\" (UniqueName: \"kubernetes.io/projected/8384bf14-e866-4e85-adaa-858bafca8741-kube-api-access-s5wxh\") pod \"openshift-config-operator-7777fb866f-ks8v2\" (UID: \"8384bf14-e866-4e85-adaa-858bafca8741\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ks8v2" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.641217 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7b99437-26b8-4ac8-876e-60b2a92f70bb-config\") pod \"console-operator-58897d9998-j5jl6\" (UID: \"b7b99437-26b8-4ac8-876e-60b2a92f70bb\") " pod="openshift-console-operator/console-operator-58897d9998-j5jl6" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.641259 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/adab756d-1eb4-4837-a93a-346a38f83df9-trusted-ca\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.641282 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.641300 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8384bf14-e866-4e85-adaa-858bafca8741-serving-cert\") pod \"openshift-config-operator-7777fb866f-ks8v2\" (UID: \"8384bf14-e866-4e85-adaa-858bafca8741\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ks8v2" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.641325 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppm42\" (UniqueName: \"kubernetes.io/projected/adab756d-1eb4-4837-a93a-346a38f83df9-kube-api-access-ppm42\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.641343 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/afe00a4b-526b-4a32-9d84-1321683d01ad-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6n52l\" (UID: \"afe00a4b-526b-4a32-9d84-1321683d01ad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6n52l" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.641431 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.641670 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dmnws\" (UniqueName: \"kubernetes.io/projected/b7b99437-26b8-4ac8-876e-60b2a92f70bb-kube-api-access-dmnws\") pod \"console-operator-58897d9998-j5jl6\" (UID: \"b7b99437-26b8-4ac8-876e-60b2a92f70bb\") " pod="openshift-console-operator/console-operator-58897d9998-j5jl6" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.641697 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4dqsh\" (UniqueName: \"kubernetes.io/projected/715578d8-876e-4dd8-9d93-71360a653604-kube-api-access-4dqsh\") pod \"migrator-59844c95c7-sfrz4\" (UID: \"715578d8-876e-4dd8-9d93-71360a653604\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sfrz4" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.641735 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7b99437-26b8-4ac8-876e-60b2a92f70bb-trusted-ca\") pod \"console-operator-58897d9998-j5jl6\" (UID: \"b7b99437-26b8-4ac8-876e-60b2a92f70bb\") " pod="openshift-console-operator/console-operator-58897d9998-j5jl6" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.641785 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.641797 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thlk5\" (UniqueName: \"kubernetes.io/projected/1ec98242-7220-4547-b699-5e97bddfbca9-kube-api-access-thlk5\") pod \"packageserver-d55dfcdfc-pqs8f\" (UID: \"1ec98242-7220-4547-b699-5e97bddfbca9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.642016 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dnkpx\" (UniqueName: \"kubernetes.io/projected/f3b2043e-942e-4a01-95ae-cf8325faf27d-kube-api-access-dnkpx\") pod \"openshift-controller-manager-operator-756b6f6bc6-rjzql\" (UID: \"f3b2043e-942e-4a01-95ae-cf8325faf27d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rjzql" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.642072 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7b99437-26b8-4ac8-876e-60b2a92f70bb-serving-cert\") pod \"console-operator-58897d9998-j5jl6\" (UID: \"b7b99437-26b8-4ac8-876e-60b2a92f70bb\") " pod="openshift-console-operator/console-operator-58897d9998-j5jl6" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.642160 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/adab756d-1eb4-4837-a93a-346a38f83df9-bound-sa-token\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.642177 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/569bec4c-7397-468e-ad4f-dd69ee0ed023-audit-dir\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.642196 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/741f2e11-ce5e-4b75-9879-6493b20f9c3f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4cv2\" (UID: \"741f2e11-ce5e-4b75-9879-6493b20f9c3f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4cv2" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.642242 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b4b2ae6-2dca-4267-b764-d6ca08e601ed-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-8n5bj\" (UID: \"8b4b2ae6-2dca-4267-b764-d6ca08e601ed\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8n5bj" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.642258 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/adab756d-1eb4-4837-a93a-346a38f83df9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.642274 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4d6br\" (UniqueName: \"kubernetes.io/projected/569bec4c-7397-468e-ad4f-dd69ee0ed023-kube-api-access-4d6br\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.642300 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.642316 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.642345 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.642360 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fb8bd\" (UniqueName: \"kubernetes.io/projected/741f2e11-ce5e-4b75-9879-6493b20f9c3f-kube-api-access-fb8bd\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4cv2\" (UID: \"741f2e11-ce5e-4b75-9879-6493b20f9c3f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4cv2" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.642382 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.642397 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/adab756d-1eb4-4837-a93a-346a38f83df9-registry-tls\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.642417 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3b2043e-942e-4a01-95ae-cf8325faf27d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rjzql\" (UID: \"f3b2043e-942e-4a01-95ae-cf8325faf27d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rjzql" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.642473 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.642761 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1ec98242-7220-4547-b699-5e97bddfbca9-apiservice-cert\") pod \"packageserver-d55dfcdfc-pqs8f\" (UID: \"1ec98242-7220-4547-b699-5e97bddfbca9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.642789 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f15a4328-85fb-49db-aed5-24a76ce8cc0e-config\") pod \"kube-controller-manager-operator-78b949d7b-sd8c7\" (UID: \"f15a4328-85fb-49db-aed5-24a76ce8cc0e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd8c7" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.642805 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/680df882-df5e-4a64-995b-3c1c35e5ec8e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-d7q9r\" (UID: \"680df882-df5e-4a64-995b-3c1c35e5ec8e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-d7q9r" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.642823 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/afe00a4b-526b-4a32-9d84-1321683d01ad-trusted-ca\") pod \"ingress-operator-5b745b69d9-6n52l\" (UID: \"afe00a4b-526b-4a32-9d84-1321683d01ad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6n52l" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.642840 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8384bf14-e866-4e85-adaa-858bafca8741-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ks8v2\" (UID: \"8384bf14-e866-4e85-adaa-858bafca8741\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ks8v2" Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.643734 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.143719386 +0000 UTC m=+137.462930055 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.644049 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/adab756d-1eb4-4837-a93a-346a38f83df9-registry-certificates\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.644075 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gqk8\" (UniqueName: \"kubernetes.io/projected/afe00a4b-526b-4a32-9d84-1321683d01ad-kube-api-access-2gqk8\") pod \"ingress-operator-5b745b69d9-6n52l\" (UID: \"afe00a4b-526b-4a32-9d84-1321683d01ad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6n52l" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.644096 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/afe00a4b-526b-4a32-9d84-1321683d01ad-metrics-tls\") pod \"ingress-operator-5b745b69d9-6n52l\" (UID: \"afe00a4b-526b-4a32-9d84-1321683d01ad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6n52l" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.644231 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/741f2e11-ce5e-4b75-9879-6493b20f9c3f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4cv2\" (UID: \"741f2e11-ce5e-4b75-9879-6493b20f9c3f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4cv2" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.644268 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3b2043e-942e-4a01-95ae-cf8325faf27d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-rjzql\" (UID: \"f3b2043e-942e-4a01-95ae-cf8325faf27d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rjzql" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.644329 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hkjz8\" (UniqueName: \"kubernetes.io/projected/661b7c44-b4e8-4c2c-a7b8-a752564f557b-kube-api-access-hkjz8\") pod \"machine-config-operator-74547568cd-5px2s\" (UID: \"661b7c44-b4e8-4c2c-a7b8-a752564f557b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5px2s" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.644345 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1ec98242-7220-4547-b699-5e97bddfbca9-tmpfs\") pod \"packageserver-d55dfcdfc-pqs8f\" (UID: \"1ec98242-7220-4547-b699-5e97bddfbca9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.644972 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b4b2ae6-2dca-4267-b764-d6ca08e601ed-config\") pod \"kube-apiserver-operator-766d6c64bb-8n5bj\" (UID: \"8b4b2ae6-2dca-4267-b764-d6ca08e601ed\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8n5bj" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.645222 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/680df882-df5e-4a64-995b-3c1c35e5ec8e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-d7q9r\" (UID: \"680df882-df5e-4a64-995b-3c1c35e5ec8e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-d7q9r" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.645365 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f15a4328-85fb-49db-aed5-24a76ce8cc0e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-sd8c7\" (UID: \"f15a4328-85fb-49db-aed5-24a76ce8cc0e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd8c7" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.648767 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/680df882-df5e-4a64-995b-3c1c35e5ec8e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-d7q9r\" (UID: \"680df882-df5e-4a64-995b-3c1c35e5ec8e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-d7q9r" Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.658317 4894 projected.go:288] Couldn't get configMap openshift-machine-api/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.658472 4894 projected.go:194] Error preparing data for projected volume kube-api-access-qj9dp for pod openshift-machine-api/machine-api-operator-5694c8668f-rvzr9: failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.661563 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/66f45f17-addc-4737-a47c-1ab34a289659-kube-api-access-qj9dp podName:66f45f17-addc-4737-a47c-1ab34a289659 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.158536004 +0000 UTC m=+137.477746663 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-qj9dp" (UniqueName: "kubernetes.io/projected/66f45f17-addc-4737-a47c-1ab34a289659-kube-api-access-qj9dp") pod "machine-api-operator-5694c8668f-rvzr9" (UID: "66f45f17-addc-4737-a47c-1ab34a289659") : failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.661649 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.675145 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-vtbc8"] Dec 09 15:34:02 crc kubenswrapper[4894]: W1209 15:34:02.678822 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod448379d9_ce78_4910_8651_541f26b80f4d.slice/crio-8771e7404ce53360f3f1afa44522e5eb6f96dd7657aae30e8a6c71c516b5aecf WatchSource:0}: Error finding container 8771e7404ce53360f3f1afa44522e5eb6f96dd7657aae30e8a6c71c516b5aecf: Status 404 returned error can't find the container with id 8771e7404ce53360f3f1afa44522e5eb6f96dd7657aae30e8a6c71c516b5aecf Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.682290 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.700746 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.720897 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.741870 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.749396 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.749671 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.249648808 +0000 UTC m=+137.568859477 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.749846 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.749889 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.749937 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cd7hx\" (UniqueName: \"kubernetes.io/projected/0847bb67-0cb7-4616-b715-29b6f12f67b5-kube-api-access-cd7hx\") pod \"marketplace-operator-79b997595-qq9k2\" (UID: \"0847bb67-0cb7-4616-b715-29b6f12f67b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.749964 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.750001 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.751206 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.251189939 +0000 UTC m=+137.570400608 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.753025 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/9c59e082-e6c9-4e92-8f39-e6f11f2d4941-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9qzp8\" (UID: \"9c59e082-e6c9-4e92-8f39-e6f11f2d4941\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qzp8" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.753101 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.753146 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s5wxh\" (UniqueName: \"kubernetes.io/projected/8384bf14-e866-4e85-adaa-858bafca8741-kube-api-access-s5wxh\") pod \"openshift-config-operator-7777fb866f-ks8v2\" (UID: \"8384bf14-e866-4e85-adaa-858bafca8741\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ks8v2" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.753183 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tkw7b\" (UniqueName: \"kubernetes.io/projected/1d071a99-3aa2-4109-a037-d05e278d1206-kube-api-access-tkw7b\") pod \"csi-hostpathplugin-l6cb5\" (UID: \"1d071a99-3aa2-4109-a037-d05e278d1206\") " pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.753206 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/adab756d-1eb4-4837-a93a-346a38f83df9-trusted-ca\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.753222 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.753238 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7b99437-26b8-4ac8-876e-60b2a92f70bb-config\") pod \"console-operator-58897d9998-j5jl6\" (UID: \"b7b99437-26b8-4ac8-876e-60b2a92f70bb\") " pod="openshift-console-operator/console-operator-58897d9998-j5jl6" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.753298 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/20c26380-64dd-4fe1-a9d5-febed69bea4e-metrics-tls\") pod \"dns-default-d5smt\" (UID: \"20c26380-64dd-4fe1-a9d5-febed69bea4e\") " pod="openshift-dns/dns-default-d5smt" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.753511 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9m2fx\" (UniqueName: \"kubernetes.io/projected/20c26380-64dd-4fe1-a9d5-febed69bea4e-kube-api-access-9m2fx\") pod \"dns-default-d5smt\" (UID: \"20c26380-64dd-4fe1-a9d5-febed69bea4e\") " pod="openshift-dns/dns-default-d5smt" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.753537 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8384bf14-e866-4e85-adaa-858bafca8741-serving-cert\") pod \"openshift-config-operator-7777fb866f-ks8v2\" (UID: \"8384bf14-e866-4e85-adaa-858bafca8741\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ks8v2" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.753590 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppm42\" (UniqueName: \"kubernetes.io/projected/adab756d-1eb4-4837-a93a-346a38f83df9-kube-api-access-ppm42\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.753723 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/afe00a4b-526b-4a32-9d84-1321683d01ad-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6n52l\" (UID: \"afe00a4b-526b-4a32-9d84-1321683d01ad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6n52l" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.754277 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.755290 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.755335 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/adab756d-1eb4-4837-a93a-346a38f83df9-trusted-ca\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.756338 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.756973 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b7b99437-26b8-4ac8-876e-60b2a92f70bb-config\") pod \"console-operator-58897d9998-j5jl6\" (UID: \"b7b99437-26b8-4ac8-876e-60b2a92f70bb\") " pod="openshift-console-operator/console-operator-58897d9998-j5jl6" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.757181 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.757405 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.757700 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gwqj\" (UniqueName: \"kubernetes.io/projected/9c59e082-e6c9-4e92-8f39-e6f11f2d4941-kube-api-access-2gwqj\") pod \"control-plane-machine-set-operator-78cbb6b69f-9qzp8\" (UID: \"9c59e082-e6c9-4e92-8f39-e6f11f2d4941\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qzp8" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.757747 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4dqsh\" (UniqueName: \"kubernetes.io/projected/715578d8-876e-4dd8-9d93-71360a653604-kube-api-access-4dqsh\") pod \"migrator-59844c95c7-sfrz4\" (UID: \"715578d8-876e-4dd8-9d93-71360a653604\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sfrz4" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.757768 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dmnws\" (UniqueName: \"kubernetes.io/projected/b7b99437-26b8-4ac8-876e-60b2a92f70bb-kube-api-access-dmnws\") pod \"console-operator-58897d9998-j5jl6\" (UID: \"b7b99437-26b8-4ac8-876e-60b2a92f70bb\") " pod="openshift-console-operator/console-operator-58897d9998-j5jl6" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.757785 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7b99437-26b8-4ac8-876e-60b2a92f70bb-trusted-ca\") pod \"console-operator-58897d9998-j5jl6\" (UID: \"b7b99437-26b8-4ac8-876e-60b2a92f70bb\") " pod="openshift-console-operator/console-operator-58897d9998-j5jl6" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.757802 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thlk5\" (UniqueName: \"kubernetes.io/projected/1ec98242-7220-4547-b699-5e97bddfbca9-kube-api-access-thlk5\") pod \"packageserver-d55dfcdfc-pqs8f\" (UID: \"1ec98242-7220-4547-b699-5e97bddfbca9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.757818 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2c4fd4ad-d670-46d1-b1ea-71ec15ad5f8b-cert\") pod \"ingress-canary-cjsx7\" (UID: \"2c4fd4ad-d670-46d1-b1ea-71ec15ad5f8b\") " pod="openshift-ingress-canary/ingress-canary-cjsx7" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.757854 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dnkpx\" (UniqueName: \"kubernetes.io/projected/f3b2043e-942e-4a01-95ae-cf8325faf27d-kube-api-access-dnkpx\") pod \"openshift-controller-manager-operator-756b6f6bc6-rjzql\" (UID: \"f3b2043e-942e-4a01-95ae-cf8325faf27d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rjzql" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.757880 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7b99437-26b8-4ac8-876e-60b2a92f70bb-serving-cert\") pod \"console-operator-58897d9998-j5jl6\" (UID: \"b7b99437-26b8-4ac8-876e-60b2a92f70bb\") " pod="openshift-console-operator/console-operator-58897d9998-j5jl6" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.757905 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zr7h6\" (UniqueName: \"kubernetes.io/projected/fbf24c93-1112-4ca1-a16b-3025aa412eee-kube-api-access-zr7h6\") pod \"machine-config-controller-84d6567774-lcnnn\" (UID: \"fbf24c93-1112-4ca1-a16b-3025aa412eee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lcnnn" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.757920 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0547d896-3107-481a-93a0-ff601f54c7c0-srv-cert\") pod \"catalog-operator-68c6474976-ktwf9\" (UID: \"0547d896-3107-481a-93a0-ff601f54c7c0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ktwf9" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.757958 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fbf24c93-1112-4ca1-a16b-3025aa412eee-proxy-tls\") pod \"machine-config-controller-84d6567774-lcnnn\" (UID: \"fbf24c93-1112-4ca1-a16b-3025aa412eee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lcnnn" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.757977 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/adab756d-1eb4-4837-a93a-346a38f83df9-bound-sa-token\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.757991 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/569bec4c-7397-468e-ad4f-dd69ee0ed023-audit-dir\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758008 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/741f2e11-ce5e-4b75-9879-6493b20f9c3f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4cv2\" (UID: \"741f2e11-ce5e-4b75-9879-6493b20f9c3f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4cv2" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758027 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct6w5\" (UniqueName: \"kubernetes.io/projected/0547d896-3107-481a-93a0-ff601f54c7c0-kube-api-access-ct6w5\") pod \"catalog-operator-68c6474976-ktwf9\" (UID: \"0547d896-3107-481a-93a0-ff601f54c7c0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ktwf9" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758046 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9mldd\" (UniqueName: \"kubernetes.io/projected/7940af9e-4cd3-4def-bdbe-c8a6ac61bb4d-kube-api-access-9mldd\") pod \"package-server-manager-789f6589d5-m8gtv\" (UID: \"7940af9e-4cd3-4def-bdbe-c8a6ac61bb4d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m8gtv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758062 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b4b2ae6-2dca-4267-b764-d6ca08e601ed-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-8n5bj\" (UID: \"8b4b2ae6-2dca-4267-b764-d6ca08e601ed\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8n5bj" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758076 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4d6br\" (UniqueName: \"kubernetes.io/projected/569bec4c-7397-468e-ad4f-dd69ee0ed023-kube-api-access-4d6br\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758095 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/adab756d-1eb4-4837-a93a-346a38f83df9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758118 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758136 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5600313-a000-42ce-94ca-898301d071fd-service-ca-bundle\") pod \"router-default-5444994796-sd4tb\" (UID: \"a5600313-a000-42ce-94ca-898301d071fd\") " pod="openshift-ingress/router-default-5444994796-sd4tb" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758153 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1d071a99-3aa2-4109-a037-d05e278d1206-socket-dir\") pod \"csi-hostpathplugin-l6cb5\" (UID: \"1d071a99-3aa2-4109-a037-d05e278d1206\") " pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758187 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758213 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwnl9\" (UniqueName: \"kubernetes.io/projected/a53e7868-4601-4aa0-b8cb-7c4413047123-kube-api-access-fwnl9\") pod \"machine-config-server-95rf5\" (UID: \"a53e7868-4601-4aa0-b8cb-7c4413047123\") " pod="openshift-machine-config-operator/machine-config-server-95rf5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758230 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4fd2af26-566f-48e7-989e-703ba77111de-signing-key\") pod \"service-ca-9c57cc56f-84l9b\" (UID: \"4fd2af26-566f-48e7-989e-703ba77111de\") " pod="openshift-service-ca/service-ca-9c57cc56f-84l9b" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758247 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758261 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a53e7868-4601-4aa0-b8cb-7c4413047123-certs\") pod \"machine-config-server-95rf5\" (UID: \"a53e7868-4601-4aa0-b8cb-7c4413047123\") " pod="openshift-machine-config-operator/machine-config-server-95rf5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758285 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fb8bd\" (UniqueName: \"kubernetes.io/projected/741f2e11-ce5e-4b75-9879-6493b20f9c3f-kube-api-access-fb8bd\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4cv2\" (UID: \"741f2e11-ce5e-4b75-9879-6493b20f9c3f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4cv2" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758321 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758357 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/adab756d-1eb4-4837-a93a-346a38f83df9-registry-tls\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758376 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3b2043e-942e-4a01-95ae-cf8325faf27d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rjzql\" (UID: \"f3b2043e-942e-4a01-95ae-cf8325faf27d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rjzql" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758404 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758420 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1d071a99-3aa2-4109-a037-d05e278d1206-registration-dir\") pod \"csi-hostpathplugin-l6cb5\" (UID: \"1d071a99-3aa2-4109-a037-d05e278d1206\") " pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758451 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20c26380-64dd-4fe1-a9d5-febed69bea4e-config-volume\") pod \"dns-default-d5smt\" (UID: \"20c26380-64dd-4fe1-a9d5-febed69bea4e\") " pod="openshift-dns/dns-default-d5smt" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758470 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1ec98242-7220-4547-b699-5e97bddfbca9-apiservice-cert\") pod \"packageserver-d55dfcdfc-pqs8f\" (UID: \"1ec98242-7220-4547-b699-5e97bddfbca9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758487 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/680df882-df5e-4a64-995b-3c1c35e5ec8e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-d7q9r\" (UID: \"680df882-df5e-4a64-995b-3c1c35e5ec8e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-d7q9r" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758504 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/afe00a4b-526b-4a32-9d84-1321683d01ad-trusted-ca\") pod \"ingress-operator-5b745b69d9-6n52l\" (UID: \"afe00a4b-526b-4a32-9d84-1321683d01ad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6n52l" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758521 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8384bf14-e866-4e85-adaa-858bafca8741-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ks8v2\" (UID: \"8384bf14-e866-4e85-adaa-858bafca8741\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ks8v2" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758522 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758537 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f15a4328-85fb-49db-aed5-24a76ce8cc0e-config\") pod \"kube-controller-manager-operator-78b949d7b-sd8c7\" (UID: \"f15a4328-85fb-49db-aed5-24a76ce8cc0e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd8c7" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758580 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/adab756d-1eb4-4837-a93a-346a38f83df9-registry-certificates\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758602 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gqk8\" (UniqueName: \"kubernetes.io/projected/afe00a4b-526b-4a32-9d84-1321683d01ad-kube-api-access-2gqk8\") pod \"ingress-operator-5b745b69d9-6n52l\" (UID: \"afe00a4b-526b-4a32-9d84-1321683d01ad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6n52l" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758622 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1d071a99-3aa2-4109-a037-d05e278d1206-mountpoint-dir\") pod \"csi-hostpathplugin-l6cb5\" (UID: \"1d071a99-3aa2-4109-a037-d05e278d1206\") " pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758658 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/afe00a4b-526b-4a32-9d84-1321683d01ad-metrics-tls\") pod \"ingress-operator-5b745b69d9-6n52l\" (UID: \"afe00a4b-526b-4a32-9d84-1321683d01ad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6n52l" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758675 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-crdp7\" (UniqueName: \"kubernetes.io/projected/a5600313-a000-42ce-94ca-898301d071fd-kube-api-access-crdp7\") pod \"router-default-5444994796-sd4tb\" (UID: \"a5600313-a000-42ce-94ca-898301d071fd\") " pod="openshift-ingress/router-default-5444994796-sd4tb" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758691 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbvzc\" (UniqueName: \"kubernetes.io/projected/f67460f9-bc45-47e0-b186-f66dea9a578c-kube-api-access-jbvzc\") pod \"collect-profiles-29421570-jcf7r\" (UID: \"f67460f9-bc45-47e0-b186-f66dea9a578c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758708 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1d071a99-3aa2-4109-a037-d05e278d1206-csi-data-dir\") pod \"csi-hostpathplugin-l6cb5\" (UID: \"1d071a99-3aa2-4109-a037-d05e278d1206\") " pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758723 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0547d896-3107-481a-93a0-ff601f54c7c0-profile-collector-cert\") pod \"catalog-operator-68c6474976-ktwf9\" (UID: \"0547d896-3107-481a-93a0-ff601f54c7c0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ktwf9" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758754 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/741f2e11-ce5e-4b75-9879-6493b20f9c3f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4cv2\" (UID: \"741f2e11-ce5e-4b75-9879-6493b20f9c3f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4cv2" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758780 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3b2043e-942e-4a01-95ae-cf8325faf27d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-rjzql\" (UID: \"f3b2043e-942e-4a01-95ae-cf8325faf27d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rjzql" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758805 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hkjz8\" (UniqueName: \"kubernetes.io/projected/661b7c44-b4e8-4c2c-a7b8-a752564f557b-kube-api-access-hkjz8\") pod \"machine-config-operator-74547568cd-5px2s\" (UID: \"661b7c44-b4e8-4c2c-a7b8-a752564f557b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5px2s" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758827 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f67460f9-bc45-47e0-b186-f66dea9a578c-config-volume\") pod \"collect-profiles-29421570-jcf7r\" (UID: \"f67460f9-bc45-47e0-b186-f66dea9a578c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758873 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1ec98242-7220-4547-b699-5e97bddfbca9-tmpfs\") pod \"packageserver-d55dfcdfc-pqs8f\" (UID: \"1ec98242-7220-4547-b699-5e97bddfbca9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758938 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b4b2ae6-2dca-4267-b764-d6ca08e601ed-config\") pod \"kube-apiserver-operator-766d6c64bb-8n5bj\" (UID: \"8b4b2ae6-2dca-4267-b764-d6ca08e601ed\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8n5bj" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758967 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f15a4328-85fb-49db-aed5-24a76ce8cc0e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-sd8c7\" (UID: \"f15a4328-85fb-49db-aed5-24a76ce8cc0e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd8c7" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758983 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/680df882-df5e-4a64-995b-3c1c35e5ec8e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-d7q9r\" (UID: \"680df882-df5e-4a64-995b-3c1c35e5ec8e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-d7q9r" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.758998 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f67460f9-bc45-47e0-b186-f66dea9a578c-secret-volume\") pod \"collect-profiles-29421570-jcf7r\" (UID: \"f67460f9-bc45-47e0-b186-f66dea9a578c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759035 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0847bb67-0cb7-4616-b715-29b6f12f67b5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qq9k2\" (UID: \"0847bb67-0cb7-4616-b715-29b6f12f67b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759053 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/680df882-df5e-4a64-995b-3c1c35e5ec8e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-d7q9r\" (UID: \"680df882-df5e-4a64-995b-3c1c35e5ec8e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-d7q9r" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759077 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7395364b-c068-4823-9591-513e1d07e0af-config\") pod \"service-ca-operator-777779d784-gr92k\" (UID: \"7395364b-c068-4823-9591-513e1d07e0af\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gr92k" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759151 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/661b7c44-b4e8-4c2c-a7b8-a752564f557b-images\") pod \"machine-config-operator-74547568cd-5px2s\" (UID: \"661b7c44-b4e8-4c2c-a7b8-a752564f557b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5px2s" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759172 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4fd2af26-566f-48e7-989e-703ba77111de-signing-cabundle\") pod \"service-ca-9c57cc56f-84l9b\" (UID: \"4fd2af26-566f-48e7-989e-703ba77111de\") " pod="openshift-service-ca/service-ca-9c57cc56f-84l9b" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759207 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f15a4328-85fb-49db-aed5-24a76ce8cc0e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-sd8c7\" (UID: \"f15a4328-85fb-49db-aed5-24a76ce8cc0e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd8c7" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759238 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rv4nb\" (UniqueName: \"kubernetes.io/projected/2c4fd4ad-d670-46d1-b1ea-71ec15ad5f8b-kube-api-access-rv4nb\") pod \"ingress-canary-cjsx7\" (UID: \"2c4fd4ad-d670-46d1-b1ea-71ec15ad5f8b\") " pod="openshift-ingress-canary/ingress-canary-cjsx7" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759256 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a53e7868-4601-4aa0-b8cb-7c4413047123-node-bootstrap-token\") pod \"machine-config-server-95rf5\" (UID: \"a53e7868-4601-4aa0-b8cb-7c4413047123\") " pod="openshift-machine-config-operator/machine-config-server-95rf5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759270 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1d071a99-3aa2-4109-a037-d05e278d1206-plugins-dir\") pod \"csi-hostpathplugin-l6cb5\" (UID: \"1d071a99-3aa2-4109-a037-d05e278d1206\") " pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759288 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vztkq\" (UniqueName: \"kubernetes.io/projected/a1be956b-214a-400f-8f5b-201fbe268090-kube-api-access-vztkq\") pod \"multus-admission-controller-857f4d67dd-7nlp9\" (UID: \"a1be956b-214a-400f-8f5b-201fbe268090\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7nlp9" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759305 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fbf24c93-1112-4ca1-a16b-3025aa412eee-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-lcnnn\" (UID: \"fbf24c93-1112-4ca1-a16b-3025aa412eee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lcnnn" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759324 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q56cg\" (UniqueName: \"kubernetes.io/projected/4fd2af26-566f-48e7-989e-703ba77111de-kube-api-access-q56cg\") pod \"service-ca-9c57cc56f-84l9b\" (UID: \"4fd2af26-566f-48e7-989e-703ba77111de\") " pod="openshift-service-ca/service-ca-9c57cc56f-84l9b" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759338 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtcpr\" (UniqueName: \"kubernetes.io/projected/7395364b-c068-4823-9591-513e1d07e0af-kube-api-access-mtcpr\") pod \"service-ca-operator-777779d784-gr92k\" (UID: \"7395364b-c068-4823-9591-513e1d07e0af\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gr92k" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759356 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-audit-policies\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759374 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7395364b-c068-4823-9591-513e1d07e0af-serving-cert\") pod \"service-ca-operator-777779d784-gr92k\" (UID: \"7395364b-c068-4823-9591-513e1d07e0af\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gr92k" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759390 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a5600313-a000-42ce-94ca-898301d071fd-metrics-certs\") pod \"router-default-5444994796-sd4tb\" (UID: \"a5600313-a000-42ce-94ca-898301d071fd\") " pod="openshift-ingress/router-default-5444994796-sd4tb" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759415 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/661b7c44-b4e8-4c2c-a7b8-a752564f557b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5px2s\" (UID: \"661b7c44-b4e8-4c2c-a7b8-a752564f557b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5px2s" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759433 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/661b7c44-b4e8-4c2c-a7b8-a752564f557b-proxy-tls\") pod \"machine-config-operator-74547568cd-5px2s\" (UID: \"661b7c44-b4e8-4c2c-a7b8-a752564f557b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5px2s" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759451 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7940af9e-4cd3-4def-bdbe-c8a6ac61bb4d-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-m8gtv\" (UID: \"7940af9e-4cd3-4def-bdbe-c8a6ac61bb4d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m8gtv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759480 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a5600313-a000-42ce-94ca-898301d071fd-default-certificate\") pod \"router-default-5444994796-sd4tb\" (UID: \"a5600313-a000-42ce-94ca-898301d071fd\") " pod="openshift-ingress/router-default-5444994796-sd4tb" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759497 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a1be956b-214a-400f-8f5b-201fbe268090-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7nlp9\" (UID: \"a1be956b-214a-400f-8f5b-201fbe268090\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7nlp9" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759527 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/adab756d-1eb4-4837-a93a-346a38f83df9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759546 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8b4b2ae6-2dca-4267-b764-d6ca08e601ed-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-8n5bj\" (UID: \"8b4b2ae6-2dca-4267-b764-d6ca08e601ed\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8n5bj" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759562 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1ec98242-7220-4547-b699-5e97bddfbca9-webhook-cert\") pod \"packageserver-d55dfcdfc-pqs8f\" (UID: \"1ec98242-7220-4547-b699-5e97bddfbca9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759590 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a5600313-a000-42ce-94ca-898301d071fd-stats-auth\") pod \"router-default-5444994796-sd4tb\" (UID: \"a5600313-a000-42ce-94ca-898301d071fd\") " pod="openshift-ingress/router-default-5444994796-sd4tb" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759606 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0847bb67-0cb7-4616-b715-29b6f12f67b5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qq9k2\" (UID: \"0847bb67-0cb7-4616-b715-29b6f12f67b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.759903 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/adab756d-1eb4-4837-a93a-346a38f83df9-registry-certificates\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.760002 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8384bf14-e866-4e85-adaa-858bafca8741-serving-cert\") pod \"openshift-config-operator-7777fb866f-ks8v2\" (UID: \"8384bf14-e866-4e85-adaa-858bafca8741\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ks8v2" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.760524 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f15a4328-85fb-49db-aed5-24a76ce8cc0e-config\") pod \"kube-controller-manager-operator-78b949d7b-sd8c7\" (UID: \"f15a4328-85fb-49db-aed5-24a76ce8cc0e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd8c7" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.761504 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3b2043e-942e-4a01-95ae-cf8325faf27d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-rjzql\" (UID: \"f3b2043e-942e-4a01-95ae-cf8325faf27d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rjzql" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.762026 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b7b99437-26b8-4ac8-876e-60b2a92f70bb-trusted-ca\") pod \"console-operator-58897d9998-j5jl6\" (UID: \"b7b99437-26b8-4ac8-876e-60b2a92f70bb\") " pod="openshift-console-operator/console-operator-58897d9998-j5jl6" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.762299 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/1ec98242-7220-4547-b699-5e97bddfbca9-tmpfs\") pod \"packageserver-d55dfcdfc-pqs8f\" (UID: \"1ec98242-7220-4547-b699-5e97bddfbca9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.762367 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.762412 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/569bec4c-7397-468e-ad4f-dd69ee0ed023-audit-dir\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.762947 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/741f2e11-ce5e-4b75-9879-6493b20f9c3f-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4cv2\" (UID: \"741f2e11-ce5e-4b75-9879-6493b20f9c3f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4cv2" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.763434 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/661b7c44-b4e8-4c2c-a7b8-a752564f557b-auth-proxy-config\") pod \"machine-config-operator-74547568cd-5px2s\" (UID: \"661b7c44-b4e8-4c2c-a7b8-a752564f557b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5px2s" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.764102 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/661b7c44-b4e8-4c2c-a7b8-a752564f557b-images\") pod \"machine-config-operator-74547568cd-5px2s\" (UID: \"661b7c44-b4e8-4c2c-a7b8-a752564f557b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5px2s" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.766480 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.766770 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3b2043e-942e-4a01-95ae-cf8325faf27d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-rjzql\" (UID: \"f3b2043e-942e-4a01-95ae-cf8325faf27d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rjzql" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.766961 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/afe00a4b-526b-4a32-9d84-1321683d01ad-trusted-ca\") pod \"ingress-operator-5b745b69d9-6n52l\" (UID: \"afe00a4b-526b-4a32-9d84-1321683d01ad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6n52l" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.766966 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.767250 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/8384bf14-e866-4e85-adaa-858bafca8741-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ks8v2\" (UID: \"8384bf14-e866-4e85-adaa-858bafca8741\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ks8v2" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.767747 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/afe00a4b-526b-4a32-9d84-1321683d01ad-metrics-tls\") pod \"ingress-operator-5b745b69d9-6n52l\" (UID: \"afe00a4b-526b-4a32-9d84-1321683d01ad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6n52l" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.768095 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.768408 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8b4b2ae6-2dca-4267-b764-d6ca08e601ed-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-8n5bj\" (UID: \"8b4b2ae6-2dca-4267-b764-d6ca08e601ed\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8n5bj" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.768912 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.770458 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8b4b2ae6-2dca-4267-b764-d6ca08e601ed-config\") pod \"kube-apiserver-operator-766d6c64bb-8n5bj\" (UID: \"8b4b2ae6-2dca-4267-b764-d6ca08e601ed\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8n5bj" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.770963 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-audit-policies\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.771718 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/680df882-df5e-4a64-995b-3c1c35e5ec8e-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-d7q9r\" (UID: \"680df882-df5e-4a64-995b-3c1c35e5ec8e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-d7q9r" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.772070 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.772082 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/741f2e11-ce5e-4b75-9879-6493b20f9c3f-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4cv2\" (UID: \"741f2e11-ce5e-4b75-9879-6493b20f9c3f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4cv2" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.772333 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/adab756d-1eb4-4837-a93a-346a38f83df9-ca-trust-extracted\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.773254 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b7b99437-26b8-4ac8-876e-60b2a92f70bb-serving-cert\") pod \"console-operator-58897d9998-j5jl6\" (UID: \"b7b99437-26b8-4ac8-876e-60b2a92f70bb\") " pod="openshift-console-operator/console-operator-58897d9998-j5jl6" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.773391 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1ec98242-7220-4547-b699-5e97bddfbca9-webhook-cert\") pod \"packageserver-d55dfcdfc-pqs8f\" (UID: \"1ec98242-7220-4547-b699-5e97bddfbca9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.773735 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/adab756d-1eb4-4837-a93a-346a38f83df9-registry-tls\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.774215 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1ec98242-7220-4547-b699-5e97bddfbca9-apiservice-cert\") pod \"packageserver-d55dfcdfc-pqs8f\" (UID: \"1ec98242-7220-4547-b699-5e97bddfbca9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.774684 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f15a4328-85fb-49db-aed5-24a76ce8cc0e-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-sd8c7\" (UID: \"f15a4328-85fb-49db-aed5-24a76ce8cc0e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd8c7" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.774693 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/661b7c44-b4e8-4c2c-a7b8-a752564f557b-proxy-tls\") pod \"machine-config-operator-74547568cd-5px2s\" (UID: \"661b7c44-b4e8-4c2c-a7b8-a752564f557b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5px2s" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.774986 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.780139 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/680df882-df5e-4a64-995b-3c1c35e5ec8e-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-d7q9r\" (UID: \"680df882-df5e-4a64-995b-3c1c35e5ec8e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-d7q9r" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.780321 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/adab756d-1eb4-4837-a93a-346a38f83df9-installation-pull-secrets\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.781477 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.802632 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.804691 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8bhhd" event={"ID":"901f250c-9e0d-4901-ac4d-0c8c336e67b8","Type":"ContainerStarted","Data":"dc5ab8ef51d965970049bcc7ab9da264ed796d9be9728774f8e42e844d7f0779"} Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.804745 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8bhhd" event={"ID":"901f250c-9e0d-4901-ac4d-0c8c336e67b8","Type":"ContainerStarted","Data":"7056bf00398cb57d6c374f90f4e0ea3a0d2b9ab109334b3c8f215b3faaff8a06"} Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.817797 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-x4lf2" event={"ID":"c83b2004-58fd-4b02-b989-9566ee5fd283","Type":"ContainerStarted","Data":"76e6b81063dd87358f36125bc2084774040d18bfe64a7668c5cd8b233b2d7f50"} Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.817983 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-x4lf2" event={"ID":"c83b2004-58fd-4b02-b989-9566ee5fd283","Type":"ContainerStarted","Data":"24916466737d38dbec1f41d0c5bd0da198218902f10940b7ed3c6e229a1458b7"} Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.819885 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-w6q9w" event={"ID":"b3748054-88ec-46d8-af2e-15565a8b772a","Type":"ContainerStarted","Data":"6cf1836b01965f8197a754cdc4bffc20008b583adc1ed8a406c8e92adb54e84d"} Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.819999 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-w6q9w" event={"ID":"b3748054-88ec-46d8-af2e-15565a8b772a","Type":"ContainerStarted","Data":"2ff7ec84129e06e1373743999368acc74c2f679c35104655afeaa221f82396ea"} Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.822326 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2thb4" event={"ID":"abf8acce-d250-4d2d-895f-3d59afee8b73","Type":"ContainerStarted","Data":"c6cc44901f9ec4de7b53fe7a186525e8fd92a089d8388a34adb08d68e48974fb"} Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.822464 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2thb4" event={"ID":"abf8acce-d250-4d2d-895f-3d59afee8b73","Type":"ContainerStarted","Data":"6cb5573e5558a18c2fede306d8e76b0631b319838533207f97bdda5cf611441e"} Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.827109 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.827997 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z94kw" event={"ID":"215f4af3-778b-4657-a9c2-dca7fd769055","Type":"ContainerStarted","Data":"1cfcdf25c1f6dfda035b111a6b22327e34a2f917d54fc68db18727b974361deb"} Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.828070 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z94kw" event={"ID":"215f4af3-778b-4657-a9c2-dca7fd769055","Type":"ContainerStarted","Data":"b9f018e03aa6feff9421f0a9699c23ab04b3076c9e81d3090aba56ee726346f1"} Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.828687 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z94kw" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.830288 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-vtbc8" event={"ID":"caed7380-e514-4f40-ae98-0be4a2fe61c3","Type":"ContainerStarted","Data":"9518dc2d03678933fc40fcc3ad60fc594e95681342d51da9b3922c53a84262be"} Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.832908 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22kg8" event={"ID":"448379d9-ce78-4910-8651-541f26b80f4d","Type":"ContainerStarted","Data":"8771e7404ce53360f3f1afa44522e5eb6f96dd7657aae30e8a6c71c516b5aecf"} Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.833358 4894 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-z94kw container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.834025 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z94kw" podUID="215f4af3-778b-4657-a9c2-dca7fd769055" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.835377 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-crhvg" event={"ID":"9d548d70-c4af-4333-8cea-f6bef5722ce5","Type":"ContainerStarted","Data":"934a294305d34c4714588f014418decc38f192e7a3074c26506ebede4edeb817"} Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.835427 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-crhvg" event={"ID":"9d548d70-c4af-4333-8cea-f6bef5722ce5","Type":"ContainerStarted","Data":"3a9d5ae85cba8e07af05cd9ed81836acb2a42edd0fafb15888d327bd0fe8f5f5"} Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.837656 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" event={"ID":"00d36531-cc42-49a2-a4c9-28f9a5ddd044","Type":"ContainerStarted","Data":"8b7241b73e8a1ff5b13e5732d0d1c9ab20a45823a3e5454d8a97537de5a9680c"} Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.837760 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" event={"ID":"00d36531-cc42-49a2-a4c9-28f9a5ddd044","Type":"ContainerStarted","Data":"adff7418b9342665f8a003a2f969034eec7663da81d7dda35af281f14d4f5cda"} Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.838245 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.840407 4894 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-qjlt6 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" start-of-body= Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.840599 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" podUID="00d36531-cc42-49a2-a4c9-28f9a5ddd044" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.10:8443/healthz\": dial tcp 10.217.0.10:8443: connect: connection refused" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.841194 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.860833 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.861240 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.3612072 +0000 UTC m=+137.680417859 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.861768 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a53e7868-4601-4aa0-b8cb-7c4413047123-certs\") pod \"machine-config-server-95rf5\" (UID: \"a53e7868-4601-4aa0-b8cb-7c4413047123\") " pod="openshift-machine-config-operator/machine-config-server-95rf5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.861911 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwnl9\" (UniqueName: \"kubernetes.io/projected/a53e7868-4601-4aa0-b8cb-7c4413047123-kube-api-access-fwnl9\") pod \"machine-config-server-95rf5\" (UID: \"a53e7868-4601-4aa0-b8cb-7c4413047123\") " pod="openshift-machine-config-operator/machine-config-server-95rf5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.862015 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4fd2af26-566f-48e7-989e-703ba77111de-signing-key\") pod \"service-ca-9c57cc56f-84l9b\" (UID: \"4fd2af26-566f-48e7-989e-703ba77111de\") " pod="openshift-service-ca/service-ca-9c57cc56f-84l9b" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.862132 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1d071a99-3aa2-4109-a037-d05e278d1206-registration-dir\") pod \"csi-hostpathplugin-l6cb5\" (UID: \"1d071a99-3aa2-4109-a037-d05e278d1206\") " pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.862243 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20c26380-64dd-4fe1-a9d5-febed69bea4e-config-volume\") pod \"dns-default-d5smt\" (UID: \"20c26380-64dd-4fe1-a9d5-febed69bea4e\") " pod="openshift-dns/dns-default-d5smt" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.865060 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1d071a99-3aa2-4109-a037-d05e278d1206-mountpoint-dir\") pod \"csi-hostpathplugin-l6cb5\" (UID: \"1d071a99-3aa2-4109-a037-d05e278d1206\") " pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.865557 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-crdp7\" (UniqueName: \"kubernetes.io/projected/a5600313-a000-42ce-94ca-898301d071fd-kube-api-access-crdp7\") pod \"router-default-5444994796-sd4tb\" (UID: \"a5600313-a000-42ce-94ca-898301d071fd\") " pod="openshift-ingress/router-default-5444994796-sd4tb" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.865812 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbvzc\" (UniqueName: \"kubernetes.io/projected/f67460f9-bc45-47e0-b186-f66dea9a578c-kube-api-access-jbvzc\") pod \"collect-profiles-29421570-jcf7r\" (UID: \"f67460f9-bc45-47e0-b186-f66dea9a578c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.866034 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1d071a99-3aa2-4109-a037-d05e278d1206-csi-data-dir\") pod \"csi-hostpathplugin-l6cb5\" (UID: \"1d071a99-3aa2-4109-a037-d05e278d1206\") " pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.866501 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0547d896-3107-481a-93a0-ff601f54c7c0-profile-collector-cert\") pod \"catalog-operator-68c6474976-ktwf9\" (UID: \"0547d896-3107-481a-93a0-ff601f54c7c0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ktwf9" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.866598 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f67460f9-bc45-47e0-b186-f66dea9a578c-config-volume\") pod \"collect-profiles-29421570-jcf7r\" (UID: \"f67460f9-bc45-47e0-b186-f66dea9a578c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.866806 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f67460f9-bc45-47e0-b186-f66dea9a578c-secret-volume\") pod \"collect-profiles-29421570-jcf7r\" (UID: \"f67460f9-bc45-47e0-b186-f66dea9a578c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.866948 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0847bb67-0cb7-4616-b715-29b6f12f67b5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qq9k2\" (UID: \"0847bb67-0cb7-4616-b715-29b6f12f67b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.867046 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7395364b-c068-4823-9591-513e1d07e0af-config\") pod \"service-ca-operator-777779d784-gr92k\" (UID: \"7395364b-c068-4823-9591-513e1d07e0af\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gr92k" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.867124 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4fd2af26-566f-48e7-989e-703ba77111de-signing-cabundle\") pod \"service-ca-9c57cc56f-84l9b\" (UID: \"4fd2af26-566f-48e7-989e-703ba77111de\") " pod="openshift-service-ca/service-ca-9c57cc56f-84l9b" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.867256 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rv4nb\" (UniqueName: \"kubernetes.io/projected/2c4fd4ad-d670-46d1-b1ea-71ec15ad5f8b-kube-api-access-rv4nb\") pod \"ingress-canary-cjsx7\" (UID: \"2c4fd4ad-d670-46d1-b1ea-71ec15ad5f8b\") " pod="openshift-ingress-canary/ingress-canary-cjsx7" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.867400 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a53e7868-4601-4aa0-b8cb-7c4413047123-node-bootstrap-token\") pod \"machine-config-server-95rf5\" (UID: \"a53e7868-4601-4aa0-b8cb-7c4413047123\") " pod="openshift-machine-config-operator/machine-config-server-95rf5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.868979 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/4fd2af26-566f-48e7-989e-703ba77111de-signing-cabundle\") pod \"service-ca-9c57cc56f-84l9b\" (UID: \"4fd2af26-566f-48e7-989e-703ba77111de\") " pod="openshift-service-ca/service-ca-9c57cc56f-84l9b" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.865435 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/1d071a99-3aa2-4109-a037-d05e278d1206-mountpoint-dir\") pod \"csi-hostpathplugin-l6cb5\" (UID: \"1d071a99-3aa2-4109-a037-d05e278d1206\") " pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.867893 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f67460f9-bc45-47e0-b186-f66dea9a578c-config-volume\") pod \"collect-profiles-29421570-jcf7r\" (UID: \"f67460f9-bc45-47e0-b186-f66dea9a578c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.868231 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7395364b-c068-4823-9591-513e1d07e0af-config\") pod \"service-ca-operator-777779d784-gr92k\" (UID: \"7395364b-c068-4823-9591-513e1d07e0af\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gr92k" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.864921 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/20c26380-64dd-4fe1-a9d5-febed69bea4e-config-volume\") pod \"dns-default-d5smt\" (UID: \"20c26380-64dd-4fe1-a9d5-febed69bea4e\") " pod="openshift-dns/dns-default-d5smt" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.866282 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/1d071a99-3aa2-4109-a037-d05e278d1206-csi-data-dir\") pod \"csi-hostpathplugin-l6cb5\" (UID: \"1d071a99-3aa2-4109-a037-d05e278d1206\") " pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.866960 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/a53e7868-4601-4aa0-b8cb-7c4413047123-certs\") pod \"machine-config-server-95rf5\" (UID: \"a53e7868-4601-4aa0-b8cb-7c4413047123\") " pod="openshift-machine-config-operator/machine-config-server-95rf5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.868985 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1d071a99-3aa2-4109-a037-d05e278d1206-plugins-dir\") pod \"csi-hostpathplugin-l6cb5\" (UID: \"1d071a99-3aa2-4109-a037-d05e278d1206\") " pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.869184 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0847bb67-0cb7-4616-b715-29b6f12f67b5-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-qq9k2\" (UID: \"0847bb67-0cb7-4616-b715-29b6f12f67b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.869208 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vztkq\" (UniqueName: \"kubernetes.io/projected/a1be956b-214a-400f-8f5b-201fbe268090-kube-api-access-vztkq\") pod \"multus-admission-controller-857f4d67dd-7nlp9\" (UID: \"a1be956b-214a-400f-8f5b-201fbe268090\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7nlp9" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.862756 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/1d071a99-3aa2-4109-a037-d05e278d1206-registration-dir\") pod \"csi-hostpathplugin-l6cb5\" (UID: \"1d071a99-3aa2-4109-a037-d05e278d1206\") " pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.869241 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fbf24c93-1112-4ca1-a16b-3025aa412eee-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-lcnnn\" (UID: \"fbf24c93-1112-4ca1-a16b-3025aa412eee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lcnnn" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.869272 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtcpr\" (UniqueName: \"kubernetes.io/projected/7395364b-c068-4823-9591-513e1d07e0af-kube-api-access-mtcpr\") pod \"service-ca-operator-777779d784-gr92k\" (UID: \"7395364b-c068-4823-9591-513e1d07e0af\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gr92k" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.869297 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q56cg\" (UniqueName: \"kubernetes.io/projected/4fd2af26-566f-48e7-989e-703ba77111de-kube-api-access-q56cg\") pod \"service-ca-9c57cc56f-84l9b\" (UID: \"4fd2af26-566f-48e7-989e-703ba77111de\") " pod="openshift-service-ca/service-ca-9c57cc56f-84l9b" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.869343 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7395364b-c068-4823-9591-513e1d07e0af-serving-cert\") pod \"service-ca-operator-777779d784-gr92k\" (UID: \"7395364b-c068-4823-9591-513e1d07e0af\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gr92k" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.869381 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a5600313-a000-42ce-94ca-898301d071fd-metrics-certs\") pod \"router-default-5444994796-sd4tb\" (UID: \"a5600313-a000-42ce-94ca-898301d071fd\") " pod="openshift-ingress/router-default-5444994796-sd4tb" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.869406 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a5600313-a000-42ce-94ca-898301d071fd-default-certificate\") pod \"router-default-5444994796-sd4tb\" (UID: \"a5600313-a000-42ce-94ca-898301d071fd\") " pod="openshift-ingress/router-default-5444994796-sd4tb" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.869429 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a1be956b-214a-400f-8f5b-201fbe268090-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7nlp9\" (UID: \"a1be956b-214a-400f-8f5b-201fbe268090\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7nlp9" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.869457 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7940af9e-4cd3-4def-bdbe-c8a6ac61bb4d-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-m8gtv\" (UID: \"7940af9e-4cd3-4def-bdbe-c8a6ac61bb4d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m8gtv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.869534 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a5600313-a000-42ce-94ca-898301d071fd-stats-auth\") pod \"router-default-5444994796-sd4tb\" (UID: \"a5600313-a000-42ce-94ca-898301d071fd\") " pod="openshift-ingress/router-default-5444994796-sd4tb" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.869558 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0847bb67-0cb7-4616-b715-29b6f12f67b5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qq9k2\" (UID: \"0847bb67-0cb7-4616-b715-29b6f12f67b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.869671 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.869702 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cd7hx\" (UniqueName: \"kubernetes.io/projected/0847bb67-0cb7-4616-b715-29b6f12f67b5-kube-api-access-cd7hx\") pod \"marketplace-operator-79b997595-qq9k2\" (UID: \"0847bb67-0cb7-4616-b715-29b6f12f67b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.869746 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/9c59e082-e6c9-4e92-8f39-e6f11f2d4941-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9qzp8\" (UID: \"9c59e082-e6c9-4e92-8f39-e6f11f2d4941\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qzp8" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.869835 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tkw7b\" (UniqueName: \"kubernetes.io/projected/1d071a99-3aa2-4109-a037-d05e278d1206-kube-api-access-tkw7b\") pod \"csi-hostpathplugin-l6cb5\" (UID: \"1d071a99-3aa2-4109-a037-d05e278d1206\") " pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.869860 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/20c26380-64dd-4fe1-a9d5-febed69bea4e-metrics-tls\") pod \"dns-default-d5smt\" (UID: \"20c26380-64dd-4fe1-a9d5-febed69bea4e\") " pod="openshift-dns/dns-default-d5smt" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.869882 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9m2fx\" (UniqueName: \"kubernetes.io/projected/20c26380-64dd-4fe1-a9d5-febed69bea4e-kube-api-access-9m2fx\") pod \"dns-default-d5smt\" (UID: \"20c26380-64dd-4fe1-a9d5-febed69bea4e\") " pod="openshift-dns/dns-default-d5smt" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.869982 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gwqj\" (UniqueName: \"kubernetes.io/projected/9c59e082-e6c9-4e92-8f39-e6f11f2d4941-kube-api-access-2gwqj\") pod \"control-plane-machine-set-operator-78cbb6b69f-9qzp8\" (UID: \"9c59e082-e6c9-4e92-8f39-e6f11f2d4941\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qzp8" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.870054 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2c4fd4ad-d670-46d1-b1ea-71ec15ad5f8b-cert\") pod \"ingress-canary-cjsx7\" (UID: \"2c4fd4ad-d670-46d1-b1ea-71ec15ad5f8b\") " pod="openshift-ingress-canary/ingress-canary-cjsx7" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.870164 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0547d896-3107-481a-93a0-ff601f54c7c0-srv-cert\") pod \"catalog-operator-68c6474976-ktwf9\" (UID: \"0547d896-3107-481a-93a0-ff601f54c7c0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ktwf9" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.870198 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zr7h6\" (UniqueName: \"kubernetes.io/projected/fbf24c93-1112-4ca1-a16b-3025aa412eee-kube-api-access-zr7h6\") pod \"machine-config-controller-84d6567774-lcnnn\" (UID: \"fbf24c93-1112-4ca1-a16b-3025aa412eee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lcnnn" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.870273 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fbf24c93-1112-4ca1-a16b-3025aa412eee-proxy-tls\") pod \"machine-config-controller-84d6567774-lcnnn\" (UID: \"fbf24c93-1112-4ca1-a16b-3025aa412eee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lcnnn" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.870329 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct6w5\" (UniqueName: \"kubernetes.io/projected/0547d896-3107-481a-93a0-ff601f54c7c0-kube-api-access-ct6w5\") pod \"catalog-operator-68c6474976-ktwf9\" (UID: \"0547d896-3107-481a-93a0-ff601f54c7c0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ktwf9" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.870356 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9mldd\" (UniqueName: \"kubernetes.io/projected/7940af9e-4cd3-4def-bdbe-c8a6ac61bb4d-kube-api-access-9mldd\") pod \"package-server-manager-789f6589d5-m8gtv\" (UID: \"7940af9e-4cd3-4def-bdbe-c8a6ac61bb4d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m8gtv" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.870394 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5600313-a000-42ce-94ca-898301d071fd-service-ca-bundle\") pod \"router-default-5444994796-sd4tb\" (UID: \"a5600313-a000-42ce-94ca-898301d071fd\") " pod="openshift-ingress/router-default-5444994796-sd4tb" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.870422 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1d071a99-3aa2-4109-a037-d05e278d1206-socket-dir\") pod \"csi-hostpathplugin-l6cb5\" (UID: \"1d071a99-3aa2-4109-a037-d05e278d1206\") " pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.870528 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/1d071a99-3aa2-4109-a037-d05e278d1206-plugins-dir\") pod \"csi-hostpathplugin-l6cb5\" (UID: \"1d071a99-3aa2-4109-a037-d05e278d1206\") " pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.870711 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f67460f9-bc45-47e0-b186-f66dea9a578c-secret-volume\") pod \"collect-profiles-29421570-jcf7r\" (UID: \"f67460f9-bc45-47e0-b186-f66dea9a578c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.872304 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/a53e7868-4601-4aa0-b8cb-7c4413047123-node-bootstrap-token\") pod \"machine-config-server-95rf5\" (UID: \"a53e7868-4601-4aa0-b8cb-7c4413047123\") " pod="openshift-machine-config-operator/machine-config-server-95rf5" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.873699 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7395364b-c068-4823-9591-513e1d07e0af-serving-cert\") pod \"service-ca-operator-777779d784-gr92k\" (UID: \"7395364b-c068-4823-9591-513e1d07e0af\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gr92k" Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.875037 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/9c59e082-e6c9-4e92-8f39-e6f11f2d4941-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-9qzp8\" (UID: \"9c59e082-e6c9-4e92-8f39-e6f11f2d4941\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qzp8" Dec 09 15:34:02 crc kubenswrapper[4894]: E1209 15:34:02.875780 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.375763541 +0000 UTC m=+137.694974200 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:02 crc kubenswrapper[4894]: I1209 15:34:02.877584 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/20c26380-64dd-4fe1-a9d5-febed69bea4e-metrics-tls\") pod \"dns-default-d5smt\" (UID: \"20c26380-64dd-4fe1-a9d5-febed69bea4e\") " pod="openshift-dns/dns-default-d5smt" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:02.882781 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fbf24c93-1112-4ca1-a16b-3025aa412eee-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-lcnnn\" (UID: \"fbf24c93-1112-4ca1-a16b-3025aa412eee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lcnnn" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:02.884179 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/a5600313-a000-42ce-94ca-898301d071fd-service-ca-bundle\") pod \"router-default-5444994796-sd4tb\" (UID: \"a5600313-a000-42ce-94ca-898301d071fd\") " pod="openshift-ingress/router-default-5444994796-sd4tb" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:02.885889 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/a5600313-a000-42ce-94ca-898301d071fd-default-certificate\") pod \"router-default-5444994796-sd4tb\" (UID: \"a5600313-a000-42ce-94ca-898301d071fd\") " pod="openshift-ingress/router-default-5444994796-sd4tb" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:02.889220 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/a5600313-a000-42ce-94ca-898301d071fd-stats-auth\") pod \"router-default-5444994796-sd4tb\" (UID: \"a5600313-a000-42ce-94ca-898301d071fd\") " pod="openshift-ingress/router-default-5444994796-sd4tb" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:02.889486 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/1d071a99-3aa2-4109-a037-d05e278d1206-socket-dir\") pod \"csi-hostpathplugin-l6cb5\" (UID: \"1d071a99-3aa2-4109-a037-d05e278d1206\") " pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:02.893441 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/0547d896-3107-481a-93a0-ff601f54c7c0-srv-cert\") pod \"catalog-operator-68c6474976-ktwf9\" (UID: \"0547d896-3107-481a-93a0-ff601f54c7c0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ktwf9" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:02.896260 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0847bb67-0cb7-4616-b715-29b6f12f67b5-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-qq9k2\" (UID: \"0847bb67-0cb7-4616-b715-29b6f12f67b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:02.909153 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/7940af9e-4cd3-4def-bdbe-c8a6ac61bb4d-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-m8gtv\" (UID: \"7940af9e-4cd3-4def-bdbe-c8a6ac61bb4d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m8gtv" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:02.971724 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:03 crc kubenswrapper[4894]: E1209 15:34:02.972346 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.472329072 +0000 UTC m=+137.791539741 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.034796 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/a1be956b-214a-400f-8f5b-201fbe268090-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-7nlp9\" (UID: \"a1be956b-214a-400f-8f5b-201fbe268090\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7nlp9" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.064988 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 09 15:34:03 crc kubenswrapper[4894]: E1209 15:34:03.065351 4894 projected.go:194] Error preparing data for projected volume kube-api-access-xq8cc for pod openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc: failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.065447 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.065477 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/a5600313-a000-42ce-94ca-898301d071fd-metrics-certs\") pod \"router-default-5444994796-sd4tb\" (UID: \"a5600313-a000-42ce-94ca-898301d071fd\") " pod="openshift-ingress/router-default-5444994796-sd4tb" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.065527 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.065680 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 09 15:34:03 crc kubenswrapper[4894]: E1209 15:34:03.065681 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f3b629b0-15d9-4e1d-908c-b21f07444427-kube-api-access-xq8cc podName:f3b629b0-15d9-4e1d-908c-b21f07444427 nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.565657435 +0000 UTC m=+137.884868104 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-xq8cc" (UniqueName: "kubernetes.io/projected/f3b629b0-15d9-4e1d-908c-b21f07444427-kube-api-access-xq8cc") pod "apiserver-7bbb656c7d-2qdjc" (UID: "f3b629b0-15d9-4e1d-908c-b21f07444427") : failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.065794 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.065820 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.065855 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.065363 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.068185 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/4fd2af26-566f-48e7-989e-703ba77111de-signing-key\") pod \"service-ca-9c57cc56f-84l9b\" (UID: \"4fd2af26-566f-48e7-989e-703ba77111de\") " pod="openshift-service-ca/service-ca-9c57cc56f-84l9b" Dec 09 15:34:03 crc kubenswrapper[4894]: E1209 15:34:03.071867 4894 projected.go:194] Error preparing data for projected volume kube-api-access-h9sjj for pod openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826: failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:03 crc kubenswrapper[4894]: E1209 15:34:03.072165 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/171c863a-36f5-48c0-be81-752684537f0b-kube-api-access-h9sjj podName:171c863a-36f5-48c0-be81-752684537f0b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.572061117 +0000 UTC m=+137.891271786 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-h9sjj" (UniqueName: "kubernetes.io/projected/171c863a-36f5-48c0-be81-752684537f0b-kube-api-access-h9sjj") pod "route-controller-manager-6576b87f9c-56826" (UID: "171c863a-36f5-48c0-be81-752684537f0b") : failed to sync configmap cache: timed out waiting for the condition Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.073668 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.075029 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/0547d896-3107-481a-93a0-ff601f54c7c0-profile-collector-cert\") pod \"catalog-operator-68c6474976-ktwf9\" (UID: \"0547d896-3107-481a-93a0-ff601f54c7c0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ktwf9" Dec 09 15:34:03 crc kubenswrapper[4894]: E1209 15:34:03.076060 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.576043124 +0000 UTC m=+137.895253863 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.078952 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fbf24c93-1112-4ca1-a16b-3025aa412eee-proxy-tls\") pod \"machine-config-controller-84d6567774-lcnnn\" (UID: \"fbf24c93-1112-4ca1-a16b-3025aa412eee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lcnnn" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.085719 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppm42\" (UniqueName: \"kubernetes.io/projected/adab756d-1eb4-4837-a93a-346a38f83df9-kube-api-access-ppm42\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.086850 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qd294\" (UniqueName: \"kubernetes.io/projected/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-kube-api-access-qd294\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.089401 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/2c4fd4ad-d670-46d1-b1ea-71ec15ad5f8b-cert\") pod \"ingress-canary-cjsx7\" (UID: \"2c4fd4ad-d670-46d1-b1ea-71ec15ad5f8b\") " pod="openshift-ingress-canary/ingress-canary-cjsx7" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.090816 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s5wxh\" (UniqueName: \"kubernetes.io/projected/8384bf14-e866-4e85-adaa-858bafca8741-kube-api-access-s5wxh\") pod \"openshift-config-operator-7777fb866f-ks8v2\" (UID: \"8384bf14-e866-4e85-adaa-858bafca8741\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ks8v2" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.096447 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/afe00a4b-526b-4a32-9d84-1321683d01ad-bound-sa-token\") pod \"ingress-operator-5b745b69d9-6n52l\" (UID: \"afe00a4b-526b-4a32-9d84-1321683d01ad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6n52l" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.114137 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-dsfnf"] Dec 09 15:34:03 crc kubenswrapper[4894]: W1209 15:34:03.122106 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod62c287c6_7ccb_4d05_8e30_c83ea71bb9d6.slice/crio-28d489ba03c6df372588e426ea7e8998811d56c6ae6aba41d5b1c250f42b8214 WatchSource:0}: Error finding container 28d489ba03c6df372588e426ea7e8998811d56c6ae6aba41d5b1c250f42b8214: Status 404 returned error can't find the container with id 28d489ba03c6df372588e426ea7e8998811d56c6ae6aba41d5b1c250f42b8214 Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.125215 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gqk8\" (UniqueName: \"kubernetes.io/projected/afe00a4b-526b-4a32-9d84-1321683d01ad-kube-api-access-2gqk8\") pod \"ingress-operator-5b745b69d9-6n52l\" (UID: \"afe00a4b-526b-4a32-9d84-1321683d01ad\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6n52l" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.143072 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4dqsh\" (UniqueName: \"kubernetes.io/projected/715578d8-876e-4dd8-9d93-71360a653604-kube-api-access-4dqsh\") pod \"migrator-59844c95c7-sfrz4\" (UID: \"715578d8-876e-4dd8-9d93-71360a653604\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sfrz4" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.156733 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dmnws\" (UniqueName: \"kubernetes.io/projected/b7b99437-26b8-4ac8-876e-60b2a92f70bb-kube-api-access-dmnws\") pod \"console-operator-58897d9998-j5jl6\" (UID: \"b7b99437-26b8-4ac8-876e-60b2a92f70bb\") " pod="openshift-console-operator/console-operator-58897d9998-j5jl6" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.177450 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thlk5\" (UniqueName: \"kubernetes.io/projected/1ec98242-7220-4547-b699-5e97bddfbca9-kube-api-access-thlk5\") pod \"packageserver-d55dfcdfc-pqs8f\" (UID: \"1ec98242-7220-4547-b699-5e97bddfbca9\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.183382 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:03 crc kubenswrapper[4894]: E1209 15:34:03.183546 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.683520997 +0000 UTC m=+138.002731666 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.184102 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qj9dp\" (UniqueName: \"kubernetes.io/projected/66f45f17-addc-4737-a47c-1ab34a289659-kube-api-access-qj9dp\") pod \"machine-api-operator-5694c8668f-rvzr9\" (UID: \"66f45f17-addc-4737-a47c-1ab34a289659\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.184267 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:03 crc kubenswrapper[4894]: E1209 15:34:03.184693 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.684677658 +0000 UTC m=+138.003888337 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.196762 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qj9dp\" (UniqueName: \"kubernetes.io/projected/66f45f17-addc-4737-a47c-1ab34a289659-kube-api-access-qj9dp\") pod \"machine-api-operator-5694c8668f-rvzr9\" (UID: \"66f45f17-addc-4737-a47c-1ab34a289659\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.201944 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hkjz8\" (UniqueName: \"kubernetes.io/projected/661b7c44-b4e8-4c2c-a7b8-a752564f557b-kube-api-access-hkjz8\") pod \"machine-config-operator-74547568cd-5px2s\" (UID: \"661b7c44-b4e8-4c2c-a7b8-a752564f557b\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5px2s" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.219237 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/adab756d-1eb4-4837-a93a-346a38f83df9-bound-sa-token\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.223352 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-j5jl6" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.233697 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ks8v2" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.244004 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6n52l" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.248371 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fb8bd\" (UniqueName: \"kubernetes.io/projected/741f2e11-ce5e-4b75-9879-6493b20f9c3f-kube-api-access-fb8bd\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4cv2\" (UID: \"741f2e11-ce5e-4b75-9879-6493b20f9c3f\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4cv2" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.263134 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/680df882-df5e-4a64-995b-3c1c35e5ec8e-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-d7q9r\" (UID: \"680df882-df5e-4a64-995b-3c1c35e5ec8e\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-d7q9r" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.287712 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:03 crc kubenswrapper[4894]: E1209 15:34:03.287966 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.787936939 +0000 UTC m=+138.107147608 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.288337 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:03 crc kubenswrapper[4894]: E1209 15:34:03.288682 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.788673159 +0000 UTC m=+138.107883818 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.307378 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4cv2" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.307552 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/f15a4328-85fb-49db-aed5-24a76ce8cc0e-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-sd8c7\" (UID: \"f15a4328-85fb-49db-aed5-24a76ce8cc0e\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd8c7" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.309740 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd8c7" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.310169 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-d7q9r" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.320342 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5px2s" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.324572 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dnkpx\" (UniqueName: \"kubernetes.io/projected/f3b2043e-942e-4a01-95ae-cf8325faf27d-kube-api-access-dnkpx\") pod \"openshift-controller-manager-operator-756b6f6bc6-rjzql\" (UID: \"f3b2043e-942e-4a01-95ae-cf8325faf27d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rjzql" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.328516 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4d6br\" (UniqueName: \"kubernetes.io/projected/569bec4c-7397-468e-ad4f-dd69ee0ed023-kube-api-access-4d6br\") pod \"oauth-openshift-558db77b4-g79bw\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.329901 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sfrz4" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.335286 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/8b4b2ae6-2dca-4267-b764-d6ca08e601ed-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-8n5bj\" (UID: \"8b4b2ae6-2dca-4267-b764-d6ca08e601ed\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8n5bj" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.340497 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.364653 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwnl9\" (UniqueName: \"kubernetes.io/projected/a53e7868-4601-4aa0-b8cb-7c4413047123-kube-api-access-fwnl9\") pod \"machine-config-server-95rf5\" (UID: \"a53e7868-4601-4aa0-b8cb-7c4413047123\") " pod="openshift-machine-config-operator/machine-config-server-95rf5" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.379777 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-crdp7\" (UniqueName: \"kubernetes.io/projected/a5600313-a000-42ce-94ca-898301d071fd-kube-api-access-crdp7\") pod \"router-default-5444994796-sd4tb\" (UID: \"a5600313-a000-42ce-94ca-898301d071fd\") " pod="openshift-ingress/router-default-5444994796-sd4tb" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.385199 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-sd4tb" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.389044 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:03 crc kubenswrapper[4894]: E1209 15:34:03.389579 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.889563486 +0000 UTC m=+138.208774155 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.395307 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbvzc\" (UniqueName: \"kubernetes.io/projected/f67460f9-bc45-47e0-b186-f66dea9a578c-kube-api-access-jbvzc\") pod \"collect-profiles-29421570-jcf7r\" (UID: \"f67460f9-bc45-47e0-b186-f66dea9a578c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.426534 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rv4nb\" (UniqueName: \"kubernetes.io/projected/2c4fd4ad-d670-46d1-b1ea-71ec15ad5f8b-kube-api-access-rv4nb\") pod \"ingress-canary-cjsx7\" (UID: \"2c4fd4ad-d670-46d1-b1ea-71ec15ad5f8b\") " pod="openshift-ingress-canary/ingress-canary-cjsx7" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.437526 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.453497 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vztkq\" (UniqueName: \"kubernetes.io/projected/a1be956b-214a-400f-8f5b-201fbe268090-kube-api-access-vztkq\") pod \"multus-admission-controller-857f4d67dd-7nlp9\" (UID: \"a1be956b-214a-400f-8f5b-201fbe268090\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-7nlp9" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.463321 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtcpr\" (UniqueName: \"kubernetes.io/projected/7395364b-c068-4823-9591-513e1d07e0af-kube-api-access-mtcpr\") pod \"service-ca-operator-777779d784-gr92k\" (UID: \"7395364b-c068-4823-9591-513e1d07e0af\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-gr92k" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.481285 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-95rf5" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.481420 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9m2fx\" (UniqueName: \"kubernetes.io/projected/20c26380-64dd-4fe1-a9d5-febed69bea4e-kube-api-access-9m2fx\") pod \"dns-default-d5smt\" (UID: \"20c26380-64dd-4fe1-a9d5-febed69bea4e\") " pod="openshift-dns/dns-default-d5smt" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.486190 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-d5smt" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.491011 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:03 crc kubenswrapper[4894]: E1209 15:34:03.491505 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:03.99149323 +0000 UTC m=+138.310703899 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.497199 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-cjsx7" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.507007 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zr7h6\" (UniqueName: \"kubernetes.io/projected/fbf24c93-1112-4ca1-a16b-3025aa412eee-kube-api-access-zr7h6\") pod \"machine-config-controller-84d6567774-lcnnn\" (UID: \"fbf24c93-1112-4ca1-a16b-3025aa412eee\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lcnnn" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.541064 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q56cg\" (UniqueName: \"kubernetes.io/projected/4fd2af26-566f-48e7-989e-703ba77111de-kube-api-access-q56cg\") pod \"service-ca-9c57cc56f-84l9b\" (UID: \"4fd2af26-566f-48e7-989e-703ba77111de\") " pod="openshift-service-ca/service-ca-9c57cc56f-84l9b" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.549115 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gwqj\" (UniqueName: \"kubernetes.io/projected/9c59e082-e6c9-4e92-8f39-e6f11f2d4941-kube-api-access-2gwqj\") pod \"control-plane-machine-set-operator-78cbb6b69f-9qzp8\" (UID: \"9c59e082-e6c9-4e92-8f39-e6f11f2d4941\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qzp8" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.551934 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8n5bj" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.559489 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.566124 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tkw7b\" (UniqueName: \"kubernetes.io/projected/1d071a99-3aa2-4109-a037-d05e278d1206-kube-api-access-tkw7b\") pod \"csi-hostpathplugin-l6cb5\" (UID: \"1d071a99-3aa2-4109-a037-d05e278d1206\") " pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.602422 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rjzql" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.602511 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.602821 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq8cc\" (UniqueName: \"kubernetes.io/projected/f3b629b0-15d9-4e1d-908c-b21f07444427-kube-api-access-xq8cc\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.602919 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9sjj\" (UniqueName: \"kubernetes.io/projected/171c863a-36f5-48c0-be81-752684537f0b-kube-api-access-h9sjj\") pod \"route-controller-manager-6576b87f9c-56826\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:34:03 crc kubenswrapper[4894]: E1209 15:34:03.603015 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:04.102995081 +0000 UTC m=+138.422205750 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.607709 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9sjj\" (UniqueName: \"kubernetes.io/projected/171c863a-36f5-48c0-be81-752684537f0b-kube-api-access-h9sjj\") pod \"route-controller-manager-6576b87f9c-56826\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.618249 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq8cc\" (UniqueName: \"kubernetes.io/projected/f3b629b0-15d9-4e1d-908c-b21f07444427-kube-api-access-xq8cc\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.631650 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9mldd\" (UniqueName: \"kubernetes.io/projected/7940af9e-4cd3-4def-bdbe-c8a6ac61bb4d-kube-api-access-9mldd\") pod \"package-server-manager-789f6589d5-m8gtv\" (UID: \"7940af9e-4cd3-4def-bdbe-c8a6ac61bb4d\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m8gtv" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.637115 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct6w5\" (UniqueName: \"kubernetes.io/projected/0547d896-3107-481a-93a0-ff601f54c7c0-kube-api-access-ct6w5\") pod \"catalog-operator-68c6474976-ktwf9\" (UID: \"0547d896-3107-481a-93a0-ff601f54c7c0\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ktwf9" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.640378 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cd7hx\" (UniqueName: \"kubernetes.io/projected/0847bb67-0cb7-4616-b715-29b6f12f67b5-kube-api-access-cd7hx\") pod \"marketplace-operator-79b997595-qq9k2\" (UID: \"0847bb67-0cb7-4616-b715-29b6f12f67b5\") " pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.660676 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ks8v2"] Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.668934 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ktwf9" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.669501 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-7nlp9" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.677567 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m8gtv" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.694125 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qzp8" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.704791 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lcnnn" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.704870 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-config\") pod \"route-controller-manager-6576b87f9c-56826\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.704898 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f3b629b0-15d9-4e1d-908c-b21f07444427-audit-policies\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.704918 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.704938 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-etcd-client\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.704972 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/66f45f17-addc-4737-a47c-1ab34a289659-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-rvzr9\" (UID: \"66f45f17-addc-4737-a47c-1ab34a289659\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.705008 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66f45f17-addc-4737-a47c-1ab34a289659-config\") pod \"machine-api-operator-5694c8668f-rvzr9\" (UID: \"66f45f17-addc-4737-a47c-1ab34a289659\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.705025 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-serving-cert\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.705053 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0d089cc1-fe1c-4ad4-a707-27c0e809649a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fmq8p\" (UID: \"0d089cc1-fe1c-4ad4-a707-27c0e809649a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fmq8p" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.705075 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/171c863a-36f5-48c0-be81-752684537f0b-serving-cert\") pod \"route-controller-manager-6576b87f9c-56826\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.705093 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-encryption-config\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.705107 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-serving-cert\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.705135 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-etcd-serving-ca\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.705175 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-client-ca\") pod \"route-controller-manager-6576b87f9c-56826\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:34:03 crc kubenswrapper[4894]: E1209 15:34:03.705754 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:04.205735838 +0000 UTC m=+138.524946507 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.705915 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/66f45f17-addc-4737-a47c-1ab34a289659-config\") pod \"machine-api-operator-5694c8668f-rvzr9\" (UID: \"66f45f17-addc-4737-a47c-1ab34a289659\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.707188 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-config\") pod \"route-controller-manager-6576b87f9c-56826\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.707891 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f3b629b0-15d9-4e1d-908c-b21f07444427-audit-policies\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.710582 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-serving-cert\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.711736 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/171c863a-36f5-48c0-be81-752684537f0b-serving-cert\") pod \"route-controller-manager-6576b87f9c-56826\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.712841 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/66f45f17-addc-4737-a47c-1ab34a289659-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-rvzr9\" (UID: \"66f45f17-addc-4737-a47c-1ab34a289659\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.713995 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-etcd-client\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.714609 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-etcd-serving-ca\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.714977 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-84l9b" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.715511 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/39bada6f-31fa-46b6-bd31-6f43cc0f6aa4-serving-cert\") pod \"apiserver-76f77b778f-bcfbr\" (UID: \"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4\") " pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.716165 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/0d089cc1-fe1c-4ad4-a707-27c0e809649a-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-fmq8p\" (UID: \"0d089cc1-fe1c-4ad4-a707-27c0e809649a\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fmq8p" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.716451 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f3b629b0-15d9-4e1d-908c-b21f07444427-encryption-config\") pod \"apiserver-7bbb656c7d-2qdjc\" (UID: \"f3b629b0-15d9-4e1d-908c-b21f07444427\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.721669 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.751239 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-client-ca\") pod \"route-controller-manager-6576b87f9c-56826\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.752423 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gr92k" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.806188 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:03 crc kubenswrapper[4894]: E1209 15:34:03.806437 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:04.306415139 +0000 UTC m=+138.625625818 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.806544 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:03 crc kubenswrapper[4894]: E1209 15:34:03.806850 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:04.306841211 +0000 UTC m=+138.626051890 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.816732 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.885379 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.901545 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.907775 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:03 crc kubenswrapper[4894]: E1209 15:34:03.908112 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:04.408090947 +0000 UTC m=+138.727301616 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.957576 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.959169 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-95rf5" event={"ID":"a53e7868-4601-4aa0-b8cb-7c4413047123","Type":"ContainerStarted","Data":"ca93bf79220257b3924add30729b6dcebf5c9e270b8f02793fcdb897b02015f4"} Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.972384 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-vtbc8" event={"ID":"caed7380-e514-4f40-ae98-0be4a2fe61c3","Type":"ContainerStarted","Data":"9baa64c69e1084e856ee400fbb406feac4de764877dcf68812c81af866bbb412"} Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.973448 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-vtbc8" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.979798 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fmq8p" Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.994298 4894 patch_prober.go:28] interesting pod/downloads-7954f5f757-vtbc8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 09 15:34:03 crc kubenswrapper[4894]: I1209 15:34:03.994354 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vtbc8" podUID="caed7380-e514-4f40-ae98-0be4a2fe61c3" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:03.996040 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.009293 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:04 crc kubenswrapper[4894]: E1209 15:34:04.009726 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:04.509714023 +0000 UTC m=+138.828924692 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.091440 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-sd4tb" event={"ID":"a5600313-a000-42ce-94ca-898301d071fd","Type":"ContainerStarted","Data":"9c20e145c2f077750779ec959f1ce2c51286165317e1fbe4fd1e3dcf80325cea"} Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.109953 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:04 crc kubenswrapper[4894]: E1209 15:34:04.110155 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:04.610124627 +0000 UTC m=+138.929335296 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.111684 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:04 crc kubenswrapper[4894]: E1209 15:34:04.112152 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:04.612138631 +0000 UTC m=+138.931349300 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.129199 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-x4lf2" event={"ID":"c83b2004-58fd-4b02-b989-9566ee5fd283","Type":"ContainerStarted","Data":"196ef4b9c36f3cd365c5a56c0a7a45a9998cc7c60648fe07790288bfc35be4ca"} Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.138984 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" event={"ID":"62c287c6-7ccb-4d05-8e30-c83ea71bb9d6","Type":"ContainerStarted","Data":"28d489ba03c6df372588e426ea7e8998811d56c6ae6aba41d5b1c250f42b8214"} Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.140591 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22kg8" event={"ID":"448379d9-ce78-4910-8651-541f26b80f4d","Type":"ContainerStarted","Data":"5722c388c75b95f3c733622b01bbbfdea86aba92708f9984af2136854d5e0421"} Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.143198 4894 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-z94kw container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" start-of-body= Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.143261 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z94kw" podUID="215f4af3-778b-4657-a9c2-dca7fd769055" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.21:8443/healthz\": dial tcp 10.217.0.21:8443: connect: connection refused" Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.165364 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.212163 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:04 crc kubenswrapper[4894]: E1209 15:34:04.212966 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:04.712938275 +0000 UTC m=+139.032148954 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.329380 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:04 crc kubenswrapper[4894]: E1209 15:34:04.329895 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:04.829880562 +0000 UTC m=+139.149091231 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.384884 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-d7q9r"] Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.388066 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4cv2"] Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.406984 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-j5jl6"] Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.431252 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:04 crc kubenswrapper[4894]: E1209 15:34:04.431464 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:04.931440817 +0000 UTC m=+139.250651486 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.431896 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:04 crc kubenswrapper[4894]: E1209 15:34:04.436442 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:04.93641355 +0000 UTC m=+139.255624209 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.534613 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:04 crc kubenswrapper[4894]: E1209 15:34:04.534794 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:05.03476736 +0000 UTC m=+139.353978029 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.535158 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:04 crc kubenswrapper[4894]: E1209 15:34:04.535519 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:05.035506249 +0000 UTC m=+139.354716908 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.637283 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:04 crc kubenswrapper[4894]: E1209 15:34:04.637771 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:05.137755392 +0000 UTC m=+139.456966061 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.679370 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z94kw" podStartSLOduration=119.679350768 podStartE2EDuration="1m59.679350768s" podCreationTimestamp="2025-12-09 15:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:04.678225628 +0000 UTC m=+138.997436297" watchObservedRunningTime="2025-12-09 15:34:04.679350768 +0000 UTC m=+138.998561457" Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.738357 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:04 crc kubenswrapper[4894]: E1209 15:34:04.738744 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:05.238731271 +0000 UTC m=+139.557941940 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.839192 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:04 crc kubenswrapper[4894]: E1209 15:34:04.839540 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:05.339510995 +0000 UTC m=+139.658721674 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.839714 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:04 crc kubenswrapper[4894]: E1209 15:34:04.840036 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:05.340026279 +0000 UTC m=+139.659237018 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.940548 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:04 crc kubenswrapper[4894]: E1209 15:34:04.940731 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:05.440699659 +0000 UTC m=+139.759910328 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.940838 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:04 crc kubenswrapper[4894]: E1209 15:34:04.941203 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:05.441177633 +0000 UTC m=+139.760388302 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:04 crc kubenswrapper[4894]: I1209 15:34:04.959667 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-8bhhd" podStartSLOduration=120.959648348 podStartE2EDuration="2m0.959648348s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:04.95784042 +0000 UTC m=+139.277051089" watchObservedRunningTime="2025-12-09 15:34:04.959648348 +0000 UTC m=+139.278859017" Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.043566 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" podStartSLOduration=121.043545059 podStartE2EDuration="2m1.043545059s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:05.004196953 +0000 UTC m=+139.323407622" watchObservedRunningTime="2025-12-09 15:34:05.043545059 +0000 UTC m=+139.362755738" Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.044378 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:05 crc kubenswrapper[4894]: E1209 15:34:05.044650 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:05.544624908 +0000 UTC m=+139.863835577 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.150518 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-22kg8" podStartSLOduration=121.150477327 podStartE2EDuration="2m1.150477327s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:05.127591974 +0000 UTC m=+139.446802643" watchObservedRunningTime="2025-12-09 15:34:05.150477327 +0000 UTC m=+139.469688006" Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.152027 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" podStartSLOduration=121.152014139 podStartE2EDuration="2m1.152014139s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:05.085533356 +0000 UTC m=+139.404744025" watchObservedRunningTime="2025-12-09 15:34:05.152014139 +0000 UTC m=+139.471224818" Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.171808 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:05 crc kubenswrapper[4894]: E1209 15:34:05.178332 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:05.678311754 +0000 UTC m=+139.997522433 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.211307 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-crhvg" podStartSLOduration=121.21129085 podStartE2EDuration="2m1.21129085s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:05.210739414 +0000 UTC m=+139.529950093" watchObservedRunningTime="2025-12-09 15:34:05.21129085 +0000 UTC m=+139.530501519" Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.261809 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-j5jl6" event={"ID":"b7b99437-26b8-4ac8-876e-60b2a92f70bb","Type":"ContainerStarted","Data":"975d7f276564a0622004a330d30095eec5ad2b2af5edaaa06495ffeb8230d2e1"} Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.280897 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4cv2" event={"ID":"741f2e11-ce5e-4b75-9879-6493b20f9c3f","Type":"ContainerStarted","Data":"90c4a53d130c12ddd5615e26e1372c693251623525072e95ceb0fca57261bc2b"} Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.286247 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sfrz4"] Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.296177 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:05 crc kubenswrapper[4894]: E1209 15:34:05.313717 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:05.813691257 +0000 UTC m=+140.132901946 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.314183 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:05 crc kubenswrapper[4894]: E1209 15:34:05.314507 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:05.814498138 +0000 UTC m=+140.133708807 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.320715 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8n5bj"] Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.327033 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-6n52l"] Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.330237 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ks8v2" event={"ID":"8384bf14-e866-4e85-adaa-858bafca8741","Type":"ContainerStarted","Data":"9d7b84e8e3483031fa27e661d2466499b503cc87189fce635200dfc864ccc2f4"} Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.335006 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd8c7"] Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.336041 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-d7q9r" event={"ID":"680df882-df5e-4a64-995b-3c1c35e5ec8e","Type":"ContainerStarted","Data":"2e098986139b2ba456ad42ddfe4b226309cbe0a04663badc328f5cae599f5028"} Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.339299 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-sd4tb" event={"ID":"a5600313-a000-42ce-94ca-898301d071fd","Type":"ContainerStarted","Data":"fe1328b22a29a06871955180b356995ac4414b83bfb946b6e2f767ec2c75dc01"} Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.344672 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-x4lf2" podStartSLOduration=121.344633087 podStartE2EDuration="2m1.344633087s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:05.294406249 +0000 UTC m=+139.613616938" watchObservedRunningTime="2025-12-09 15:34:05.344633087 +0000 UTC m=+139.663843756" Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.346716 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f"] Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.389655 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-sd4tb" Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.391997 4894 patch_prober.go:28] interesting pod/router-default-5444994796-sd4tb container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.392049 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sd4tb" podUID="a5600313-a000-42ce-94ca-898301d071fd" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.400426 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-dsfnf" event={"ID":"62c287c6-7ccb-4d05-8e30-c83ea71bb9d6","Type":"ContainerStarted","Data":"27ae74388071853b920c074564ed401f27a4bee9d371b333c2ecd14586b25444"} Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.402820 4894 patch_prober.go:28] interesting pod/downloads-7954f5f757-vtbc8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.402887 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vtbc8" podUID="caed7380-e514-4f40-ae98-0be4a2fe61c3" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.410246 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-vtbc8" podStartSLOduration=121.410223177 podStartE2EDuration="2m1.410223177s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:05.407935025 +0000 UTC m=+139.727145704" watchObservedRunningTime="2025-12-09 15:34:05.410223177 +0000 UTC m=+139.729433856" Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.417402 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:05 crc kubenswrapper[4894]: E1209 15:34:05.421261 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:05.921229732 +0000 UTC m=+140.240440461 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.421567 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-z94kw" Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.533397 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:05 crc kubenswrapper[4894]: E1209 15:34:05.536005 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:06.035986851 +0000 UTC m=+140.355197610 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.544460 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-7nlp9"] Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.571033 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r"] Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.606715 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-5px2s"] Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.622131 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-cjsx7"] Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.643497 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:05 crc kubenswrapper[4894]: E1209 15:34:05.644008 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:06.143987248 +0000 UTC m=+140.463197917 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.663549 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ktwf9"] Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.702715 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rjzql"] Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.717213 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qzp8"] Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.719918 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qq9k2"] Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.720304 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-2thb4" podStartSLOduration=121.720293625 podStartE2EDuration="2m1.720293625s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:05.680250371 +0000 UTC m=+139.999461040" watchObservedRunningTime="2025-12-09 15:34:05.720293625 +0000 UTC m=+140.039504294" Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.739202 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-d5smt"] Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.741505 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-bcfbr"] Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.747612 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:05 crc kubenswrapper[4894]: E1209 15:34:05.747874 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:06.247863905 +0000 UTC m=+140.567074574 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.751912 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-g79bw"] Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.771117 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-w6q9w" podStartSLOduration=121.771093458 podStartE2EDuration="2m1.771093458s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:05.711554951 +0000 UTC m=+140.030765620" watchObservedRunningTime="2025-12-09 15:34:05.771093458 +0000 UTC m=+140.090304127" Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.794080 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-sd4tb" podStartSLOduration=121.794063444 podStartE2EDuration="2m1.794063444s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:05.790101308 +0000 UTC m=+140.109311977" watchObservedRunningTime="2025-12-09 15:34:05.794063444 +0000 UTC m=+140.113274103" Dec 09 15:34:05 crc kubenswrapper[4894]: W1209 15:34:05.806733 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9c59e082_e6c9_4e92_8f39_e6f11f2d4941.slice/crio-fc08411e83961d0255f38f3f125c95bff2444f550aea892c7ca15e147e9cff04 WatchSource:0}: Error finding container fc08411e83961d0255f38f3f125c95bff2444f550aea892c7ca15e147e9cff04: Status 404 returned error can't find the container with id fc08411e83961d0255f38f3f125c95bff2444f550aea892c7ca15e147e9cff04 Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.851790 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:05 crc kubenswrapper[4894]: E1209 15:34:05.852166 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:06.352146442 +0000 UTC m=+140.671357111 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.883705 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m8gtv"] Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.913542 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-l6cb5"] Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.955486 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:05 crc kubenswrapper[4894]: E1209 15:34:05.955900 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:06.455885936 +0000 UTC m=+140.775096615 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:05 crc kubenswrapper[4894]: I1209 15:34:05.979334 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fmq8p"] Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.009589 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826"] Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.026739 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-lcnnn"] Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.029399 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-rvzr9"] Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.043733 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-gr92k"] Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.058462 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:06 crc kubenswrapper[4894]: E1209 15:34:06.059320 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:06.55929627 +0000 UTC m=+140.878506939 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.076891 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc"] Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.077941 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-84l9b"] Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.160468 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:06 crc kubenswrapper[4894]: E1209 15:34:06.161831 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:06.661815841 +0000 UTC m=+140.981026510 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.269443 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:06 crc kubenswrapper[4894]: E1209 15:34:06.269773 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:06.769750366 +0000 UTC m=+141.088961045 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.273773 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:06 crc kubenswrapper[4894]: E1209 15:34:06.274420 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:06.774405051 +0000 UTC m=+141.093615730 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.376507 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:06 crc kubenswrapper[4894]: E1209 15:34:06.377101 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:06.877084155 +0000 UTC m=+141.196294824 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.388249 4894 patch_prober.go:28] interesting pod/router-default-5444994796-sd4tb container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.388287 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sd4tb" podUID="a5600313-a000-42ce-94ca-898301d071fd" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.457160 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" event={"ID":"171c863a-36f5-48c0-be81-752684537f0b","Type":"ContainerStarted","Data":"0e757050ff3deaa03ba4dde7fb53c2f722babb0539f85f88aee92a279f5ec20c"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.492495 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:06 crc kubenswrapper[4894]: E1209 15:34:06.492946 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:06.992930824 +0000 UTC m=+141.312141493 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.520128 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-d7q9r" event={"ID":"680df882-df5e-4a64-995b-3c1c35e5ec8e","Type":"ContainerStarted","Data":"185b2d72a8337590c20a3651d581d0c75d04fc97073fba1334e3245b35eff99e"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.530630 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-95rf5" event={"ID":"a53e7868-4601-4aa0-b8cb-7c4413047123","Type":"ContainerStarted","Data":"f72bf3999007e19022dfb3c9e991b6787868b39b43d8c1d98eaca02c78a760b1"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.577631 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" event={"ID":"569bec4c-7397-468e-ad4f-dd69ee0ed023","Type":"ContainerStarted","Data":"e2bb18758cf504aa6fb0bb4f0e1e9df3b2409e25166b6ef20276127e03b6d511"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.594767 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:06 crc kubenswrapper[4894]: E1209 15:34:06.595264 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:07.095245258 +0000 UTC m=+141.414455927 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.598913 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-d5smt" event={"ID":"20c26380-64dd-4fe1-a9d5-febed69bea4e","Type":"ContainerStarted","Data":"5b5a7a1a12dd07b88ffe384ca6b884aff819a66b55ca4a485400d07acf64aa69"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.612349 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4cv2" event={"ID":"741f2e11-ce5e-4b75-9879-6493b20f9c3f","Type":"ContainerStarted","Data":"973c3a7603051553b05c067d0e512c072015b7e2414c42a43f9a0929bf19a901"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.625688 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qzp8" event={"ID":"9c59e082-e6c9-4e92-8f39-e6f11f2d4941","Type":"ContainerStarted","Data":"fc08411e83961d0255f38f3f125c95bff2444f550aea892c7ca15e147e9cff04"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.630543 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sfrz4" event={"ID":"715578d8-876e-4dd8-9d93-71360a653604","Type":"ContainerStarted","Data":"9a2e096aa4c909766b82b1ab1f7e77539b34ef5d640c0a58c32b66a91cdfb7b4"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.630597 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sfrz4" event={"ID":"715578d8-876e-4dd8-9d93-71360a653604","Type":"ContainerStarted","Data":"e72fdaf842f7a74d883160d5e3b92026f1c624a4a914ec949ae645bb819f15f3"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.672362 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f" event={"ID":"1ec98242-7220-4547-b699-5e97bddfbca9","Type":"ContainerStarted","Data":"b8797f96cf056eb26ae330b084250bfcf3607bd81d7db3c7c0066ee0133f6b5b"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.672404 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f" event={"ID":"1ec98242-7220-4547-b699-5e97bddfbca9","Type":"ContainerStarted","Data":"24abd309f71fd6a32e9785378d08c9dda19869a83424b7aa6434f16b301fb9d9"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.673065 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f" Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.696682 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:06 crc kubenswrapper[4894]: E1209 15:34:06.698151 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:07.198136889 +0000 UTC m=+141.517347558 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.712850 4894 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-pqs8f container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:5443/healthz\": dial tcp 10.217.0.25:5443: connect: connection refused" start-of-body= Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.712913 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f" podUID="1ec98242-7220-4547-b699-5e97bddfbca9" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.25:5443/healthz\": dial tcp 10.217.0.25:5443: connect: connection refused" Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.734222 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5px2s" event={"ID":"661b7c44-b4e8-4c2c-a7b8-a752564f557b","Type":"ContainerStarted","Data":"a0fff1a7e8751949531281097e56fa09315fa204601b6d384a0954b49d442058"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.735784 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" event={"ID":"f3b629b0-15d9-4e1d-908c-b21f07444427","Type":"ContainerStarted","Data":"525e492d1c4f87e1288d8ec639e31dcbb2ae84dbadf2c432f529b6f0077771d0"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.738489 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8n5bj" event={"ID":"8b4b2ae6-2dca-4267-b764-d6ca08e601ed","Type":"ContainerStarted","Data":"a504629446cdb5330e1a15d5893a96c7b37edc8bf8bd3a700c089fc56e60e176"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.738547 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8n5bj" event={"ID":"8b4b2ae6-2dca-4267-b764-d6ca08e601ed","Type":"ContainerStarted","Data":"b4ed68bf15d7906b8a6e1f67affdd5a4eefa5ae8c22dc86fcafd429d63162372"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.797990 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:06 crc kubenswrapper[4894]: E1209 15:34:06.799261 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:07.299242922 +0000 UTC m=+141.618453601 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.800835 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" event={"ID":"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4","Type":"ContainerStarted","Data":"ce4a9760fac96bc70aa3cfa6e42793b1a8fc46764dec568543208ca2e82f77c2"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.853923 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m8gtv" event={"ID":"7940af9e-4cd3-4def-bdbe-c8a6ac61bb4d","Type":"ContainerStarted","Data":"4fcbb21f42b63887c9398e5c5e2c37e96c755f23f6cd02353a551d7425f0d90e"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.863760 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" event={"ID":"1d071a99-3aa2-4109-a037-d05e278d1206","Type":"ContainerStarted","Data":"db94ad423b9667dde84de7eeaf264ac7f50bec0250c719b608b8bcc1f8b0693d"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.864793 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fmq8p" event={"ID":"0d089cc1-fe1c-4ad4-a707-27c0e809649a","Type":"ContainerStarted","Data":"5be0d134cf71985811f768a78f04c35d84e6394e7353e4d94d22e84d59e38e62"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.865984 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ktwf9" event={"ID":"0547d896-3107-481a-93a0-ff601f54c7c0","Type":"ContainerStarted","Data":"86b6977ea95fc234c6abc3cfe9e5a7afd2e6dacb3f7add1c68eb14ed688ebf4c"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.867145 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ktwf9" Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.889790 4894 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-ktwf9 container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" start-of-body= Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.889849 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ktwf9" podUID="0547d896-3107-481a-93a0-ff601f54c7c0" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.20:8443/healthz\": dial tcp 10.217.0.20:8443: connect: connection refused" Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.898127 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" event={"ID":"66f45f17-addc-4737-a47c-1ab34a289659","Type":"ContainerStarted","Data":"ee602ee719701138eb69364637cd65af611c78c897d32c335fc55aaf299f7f04"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.899605 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:06 crc kubenswrapper[4894]: E1209 15:34:06.899880 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:07.399870562 +0000 UTC m=+141.719081231 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.925444 4894 generic.go:334] "Generic (PLEG): container finished" podID="8384bf14-e866-4e85-adaa-858bafca8741" containerID="fe773ee407853f39c7780f700929c8f2f642bc05d08ae5272e39ad5841e06bba" exitCode=0 Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.925512 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ks8v2" event={"ID":"8384bf14-e866-4e85-adaa-858bafca8741","Type":"ContainerDied","Data":"fe773ee407853f39c7780f700929c8f2f642bc05d08ae5272e39ad5841e06bba"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.944205 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r" event={"ID":"f67460f9-bc45-47e0-b186-f66dea9a578c","Type":"ContainerStarted","Data":"8d0fe7df17a05f02f99d71e792190b18f61f1493f9d1a5f9b393d19d743f0d6b"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.972475 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-j5jl6" event={"ID":"b7b99437-26b8-4ac8-876e-60b2a92f70bb","Type":"ContainerStarted","Data":"f2c8bca9a67647d09a21218ea997e5d3ee16d034f66d5cc1e6255d2025aaf569"} Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.973469 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-j5jl6" Dec 09 15:34:06 crc kubenswrapper[4894]: I1209 15:34:06.987044 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-cjsx7" event={"ID":"2c4fd4ad-d670-46d1-b1ea-71ec15ad5f8b","Type":"ContainerStarted","Data":"bdbed949d111e30aacad9574ce61ea14a2c8538bd60c6ae50fc5ac59aeb3c538"} Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.005471 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:07 crc kubenswrapper[4894]: E1209 15:34:07.006363 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:07.506335698 +0000 UTC m=+141.825546407 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.012889 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd8c7" event={"ID":"f15a4328-85fb-49db-aed5-24a76ce8cc0e","Type":"ContainerStarted","Data":"4ace912b3c36fa256f76ac62553c0dad7440c6fbebae719a316a3e258f5b2196"} Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.013857 4894 patch_prober.go:28] interesting pod/console-operator-58897d9998-j5jl6 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.013976 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-j5jl6" podUID="b7b99437-26b8-4ac8-876e-60b2a92f70bb" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.038780 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gr92k" event={"ID":"7395364b-c068-4823-9591-513e1d07e0af","Type":"ContainerStarted","Data":"c778f166fd76b9685f24bcca59633b13296383604b4d7175987b58cccc4f6e48"} Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.063031 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rjzql" event={"ID":"f3b2043e-942e-4a01-95ae-cf8325faf27d","Type":"ContainerStarted","Data":"13c95317b0be6ae1c24373e3109979cca710ac0245aa61b81880ab5c8e516dc6"} Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.076733 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6n52l" event={"ID":"afe00a4b-526b-4a32-9d84-1321683d01ad","Type":"ContainerStarted","Data":"b8da89c935c805221c89cb65b1ca763db7b32be8a323bc0402d259e90c4402ce"} Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.076781 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6n52l" event={"ID":"afe00a4b-526b-4a32-9d84-1321683d01ad","Type":"ContainerStarted","Data":"1ad44044aed568c610c873e29d6ede0cb1317522c2753250c60bd6ba28b50969"} Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.108468 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:07 crc kubenswrapper[4894]: E1209 15:34:07.110039 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:07.61002666 +0000 UTC m=+141.929237329 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.192171 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7nlp9" event={"ID":"a1be956b-214a-400f-8f5b-201fbe268090","Type":"ContainerStarted","Data":"90004a10dd3ce5908c9a869e41842ae7a44178370824c382eb79a14643651b1a"} Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.199811 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" event={"ID":"0847bb67-0cb7-4616-b715-29b6f12f67b5","Type":"ContainerStarted","Data":"c3cf02ed9dedfecf73ef20da80c3e0c96e5f3e9f1f14f203c7041cc1cc97736d"} Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.201061 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.209727 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:07 crc kubenswrapper[4894]: E1209 15:34:07.214428 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:07.714399859 +0000 UTC m=+142.033610528 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.215388 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lcnnn" event={"ID":"fbf24c93-1112-4ca1-a16b-3025aa412eee","Type":"ContainerStarted","Data":"05f9c7b775445333ab4684184b95e3a387445bfeca4523ad20dc7b38a330ff95"} Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.218232 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:07 crc kubenswrapper[4894]: E1209 15:34:07.219694 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:07.719680591 +0000 UTC m=+142.038891260 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.220580 4894 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-qq9k2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.220648 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" podUID="0847bb67-0cb7-4616-b715-29b6f12f67b5" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.269760 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-84l9b" event={"ID":"4fd2af26-566f-48e7-989e-703ba77111de","Type":"ContainerStarted","Data":"b751fa6dd56e2f2b9c5753edd5e64072dd7b489340c9cfb93a03ea3749f4195f"} Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.277223 4894 patch_prober.go:28] interesting pod/downloads-7954f5f757-vtbc8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.277491 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vtbc8" podUID="caed7380-e514-4f40-ae98-0be4a2fe61c3" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.321206 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:07 crc kubenswrapper[4894]: E1209 15:34:07.322756 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:07.822733396 +0000 UTC m=+142.141944075 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.397744 4894 patch_prober.go:28] interesting pod/router-default-5444994796-sd4tb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 15:34:07 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 09 15:34:07 crc kubenswrapper[4894]: [+]process-running ok Dec 09 15:34:07 crc kubenswrapper[4894]: healthz check failed Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.397809 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sd4tb" podUID="a5600313-a000-42ce-94ca-898301d071fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.433293 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:07 crc kubenswrapper[4894]: E1209 15:34:07.433744 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:07.933726124 +0000 UTC m=+142.252936793 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.553599 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:07 crc kubenswrapper[4894]: E1209 15:34:07.554298 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:08.054281418 +0000 UTC m=+142.373492087 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.647121 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ktwf9" podStartSLOduration=122.647097678 podStartE2EDuration="2m2.647097678s" podCreationTimestamp="2025-12-09 15:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:07.644322333 +0000 UTC m=+141.963533002" watchObservedRunningTime="2025-12-09 15:34:07.647097678 +0000 UTC m=+141.966308347" Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.655724 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:07 crc kubenswrapper[4894]: E1209 15:34:07.655990 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:08.155979557 +0000 UTC m=+142.475190226 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.763522 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:07 crc kubenswrapper[4894]: E1209 15:34:07.763967 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:08.263953843 +0000 UTC m=+142.583164512 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.769513 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rjzql" podStartSLOduration=123.769496532 podStartE2EDuration="2m3.769496532s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:07.768165896 +0000 UTC m=+142.087376565" watchObservedRunningTime="2025-12-09 15:34:07.769496532 +0000 UTC m=+142.088707201" Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.770472 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4cv2" podStartSLOduration=122.770464408 podStartE2EDuration="2m2.770464408s" podCreationTimestamp="2025-12-09 15:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:07.705755742 +0000 UTC m=+142.024966411" watchObservedRunningTime="2025-12-09 15:34:07.770464408 +0000 UTC m=+142.089675077" Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.865035 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:07 crc kubenswrapper[4894]: E1209 15:34:07.865747 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:08.365733504 +0000 UTC m=+142.684944173 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.866140 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-d7q9r" podStartSLOduration=122.866128405 podStartE2EDuration="2m2.866128405s" podCreationTimestamp="2025-12-09 15:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:07.866058293 +0000 UTC m=+142.185268962" watchObservedRunningTime="2025-12-09 15:34:07.866128405 +0000 UTC m=+142.185339084" Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.920998 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f" podStartSLOduration=122.920976996 podStartE2EDuration="2m2.920976996s" podCreationTimestamp="2025-12-09 15:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:07.914818841 +0000 UTC m=+142.234029500" watchObservedRunningTime="2025-12-09 15:34:07.920976996 +0000 UTC m=+142.240187665" Dec 09 15:34:07 crc kubenswrapper[4894]: I1209 15:34:07.967818 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:07 crc kubenswrapper[4894]: E1209 15:34:07.968275 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:08.468257995 +0000 UTC m=+142.787468664 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.031107 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-j5jl6" podStartSLOduration=124.03108115 podStartE2EDuration="2m4.03108115s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:07.972374065 +0000 UTC m=+142.291584734" watchObservedRunningTime="2025-12-09 15:34:08.03108115 +0000 UTC m=+142.350291829" Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.072122 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:08 crc kubenswrapper[4894]: E1209 15:34:08.072503 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:08.57248675 +0000 UTC m=+142.891697419 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.090144 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" podStartSLOduration=123.090120004 podStartE2EDuration="2m3.090120004s" podCreationTimestamp="2025-12-09 15:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:08.031008459 +0000 UTC m=+142.350219138" watchObservedRunningTime="2025-12-09 15:34:08.090120004 +0000 UTC m=+142.409330683" Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.155655 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd8c7" podStartSLOduration=123.155624871 podStartE2EDuration="2m3.155624871s" podCreationTimestamp="2025-12-09 15:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:08.141937674 +0000 UTC m=+142.461148333" watchObservedRunningTime="2025-12-09 15:34:08.155624871 +0000 UTC m=+142.474835540" Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.155780 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-95rf5" podStartSLOduration=9.155775315 podStartE2EDuration="9.155775315s" podCreationTimestamp="2025-12-09 15:33:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:08.09258172 +0000 UTC m=+142.411792389" watchObservedRunningTime="2025-12-09 15:34:08.155775315 +0000 UTC m=+142.474985984" Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.175322 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:08 crc kubenswrapper[4894]: E1209 15:34:08.175980 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:08.675932446 +0000 UTC m=+142.995143115 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.176067 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:08 crc kubenswrapper[4894]: E1209 15:34:08.176536 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:08.676526252 +0000 UTC m=+142.995736921 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.244338 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-8n5bj" podStartSLOduration=124.244311801 podStartE2EDuration="2m4.244311801s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:08.232700109 +0000 UTC m=+142.551910778" watchObservedRunningTime="2025-12-09 15:34:08.244311801 +0000 UTC m=+142.563522470" Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.275066 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-cjsx7" podStartSLOduration=8.275046695 podStartE2EDuration="8.275046695s" podCreationTimestamp="2025-12-09 15:34:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:08.274924602 +0000 UTC m=+142.594135271" watchObservedRunningTime="2025-12-09 15:34:08.275046695 +0000 UTC m=+142.594257364" Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.278560 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:08 crc kubenswrapper[4894]: E1209 15:34:08.278875 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:08.778847447 +0000 UTC m=+143.098058116 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.278981 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:08 crc kubenswrapper[4894]: E1209 15:34:08.280321 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:08.780305787 +0000 UTC m=+143.099516456 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.292708 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m8gtv" event={"ID":"7940af9e-4cd3-4def-bdbe-c8a6ac61bb4d","Type":"ContainerStarted","Data":"d8acad4b2f21db708cd62d5eff595cd6b2acba29625c56750ab074e4e6a42e09"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.293820 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gr92k" event={"ID":"7395364b-c068-4823-9591-513e1d07e0af","Type":"ContainerStarted","Data":"7c73a2c20b70cec824c314bab82518618a04a6a506e0223ffb228dc8ea5d73f4"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.297942 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qzp8" event={"ID":"9c59e082-e6c9-4e92-8f39-e6f11f2d4941","Type":"ContainerStarted","Data":"ca013abb9fd19a34d4f068c0be8e7834a1072d81898062bbaaa7a5c9fa8507c9"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.299401 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-sd8c7" event={"ID":"f15a4328-85fb-49db-aed5-24a76ce8cc0e","Type":"ContainerStarted","Data":"fd85a4c8b32fa6a045ecaec3a963a711b5258cc37e69c110e74865086a786753"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.303211 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5px2s" event={"ID":"661b7c44-b4e8-4c2c-a7b8-a752564f557b","Type":"ContainerStarted","Data":"b3762cfce17640ec1eb4a3a57c73f569963a806c7e926c81e02c4d78af5e24e7"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.303251 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5px2s" event={"ID":"661b7c44-b4e8-4c2c-a7b8-a752564f557b","Type":"ContainerStarted","Data":"0b011b6b0802520e477eaf1931b61acceab29006e796c3afdc367089ed1e0c7f"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.319946 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-84l9b" event={"ID":"4fd2af26-566f-48e7-989e-703ba77111de","Type":"ContainerStarted","Data":"b8923e324829cb41a207b50f9f0ecfb432707476e8a924663c8c31709c4503b2"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.383134 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-d5smt" event={"ID":"20c26380-64dd-4fe1-a9d5-febed69bea4e","Type":"ContainerStarted","Data":"214eddead8b199d2b68a723e1855fe44414a579f80a07f3da84444b9dccb071f"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.383568 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:08 crc kubenswrapper[4894]: E1209 15:34:08.385243 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:08.885223371 +0000 UTC m=+143.204434040 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.409920 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" event={"ID":"171c863a-36f5-48c0-be81-752684537f0b","Type":"ContainerStarted","Data":"79966100b093d6f879432f696bf6008f04634486cdb672b037d179f09313946f"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.411673 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-9qzp8" podStartSLOduration=123.41165497 podStartE2EDuration="2m3.41165497s" podCreationTimestamp="2025-12-09 15:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:08.366002755 +0000 UTC m=+142.685213434" watchObservedRunningTime="2025-12-09 15:34:08.41165497 +0000 UTC m=+142.730865639" Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.411776 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.412260 4894 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-56826 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.412421 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" podUID="171c863a-36f5-48c0-be81-752684537f0b" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.412892 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-gr92k" podStartSLOduration=123.412887483 podStartE2EDuration="2m3.412887483s" podCreationTimestamp="2025-12-09 15:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:08.400784889 +0000 UTC m=+142.719995558" watchObservedRunningTime="2025-12-09 15:34:08.412887483 +0000 UTC m=+142.732098152" Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.414010 4894 patch_prober.go:28] interesting pod/router-default-5444994796-sd4tb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 15:34:08 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 09 15:34:08 crc kubenswrapper[4894]: [+]process-running ok Dec 09 15:34:08 crc kubenswrapper[4894]: healthz check failed Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.414239 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sd4tb" podUID="a5600313-a000-42ce-94ca-898301d071fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.446386 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" event={"ID":"66f45f17-addc-4737-a47c-1ab34a289659","Type":"ContainerStarted","Data":"9eb449a6801a993f5371d719e02299f24e469825bc22e0d2fedf045d45dda9cb"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.492866 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sfrz4" event={"ID":"715578d8-876e-4dd8-9d93-71360a653604","Type":"ContainerStarted","Data":"d403afe9c4b8492b3b16548990863242db62ec7064725d8f087f6334d94cb495"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.493177 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:08 crc kubenswrapper[4894]: E1209 15:34:08.493859 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:08.993844185 +0000 UTC m=+143.313054854 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.496131 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r" event={"ID":"f67460f9-bc45-47e0-b186-f66dea9a578c","Type":"ContainerStarted","Data":"d41ad5679ea1a20ef7453e9a31ff71dd7d5f8395803d542aa6e12f46dafa93f3"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.502443 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" event={"ID":"569bec4c-7397-468e-ad4f-dd69ee0ed023","Type":"ContainerStarted","Data":"4e6e6eff4898e2307b04ddaf3a855e22814c12e1d72d16d3c33d43290e05907f"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.503117 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.503949 4894 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-g79bw container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.17:6443/healthz\": dial tcp 10.217.0.17:6443: connect: connection refused" start-of-body= Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.504053 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" podUID="569bec4c-7397-468e-ad4f-dd69ee0ed023" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.17:6443/healthz\": dial tcp 10.217.0.17:6443: connect: connection refused" Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.504084 4894 generic.go:334] "Generic (PLEG): container finished" podID="f3b629b0-15d9-4e1d-908c-b21f07444427" containerID="8aa22a13ef313a0f0fa86a6c24fd4f379b04460053ff76b6e39bcc152fc3ae18" exitCode=0 Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.504096 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" event={"ID":"f3b629b0-15d9-4e1d-908c-b21f07444427","Type":"ContainerDied","Data":"8aa22a13ef313a0f0fa86a6c24fd4f379b04460053ff76b6e39bcc152fc3ae18"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.515021 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ktwf9" event={"ID":"0547d896-3107-481a-93a0-ff601f54c7c0","Type":"ContainerStarted","Data":"6c87ce28dc15554557a345d3dbed579f3e5372aa6335dbce1185c21f0015f9b4"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.524443 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" podStartSLOduration=123.524417175 podStartE2EDuration="2m3.524417175s" podCreationTimestamp="2025-12-09 15:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:08.436451975 +0000 UTC m=+142.755662664" watchObservedRunningTime="2025-12-09 15:34:08.524417175 +0000 UTC m=+142.843627844" Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.525081 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sfrz4" podStartSLOduration=123.525072512 podStartE2EDuration="2m3.525072512s" podCreationTimestamp="2025-12-09 15:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:08.519355979 +0000 UTC m=+142.838566648" watchObservedRunningTime="2025-12-09 15:34:08.525072512 +0000 UTC m=+142.844283181" Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.537312 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7nlp9" event={"ID":"a1be956b-214a-400f-8f5b-201fbe268090","Type":"ContainerStarted","Data":"55ea6a41e1513d35a66cdfc98777f430a2457aee1fe086f3bfc6047dda315c77"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.544219 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fmq8p" event={"ID":"0d089cc1-fe1c-4ad4-a707-27c0e809649a","Type":"ContainerStarted","Data":"9457ed8db30cf58e00e21c484e75c1e7ead94a615b50ec0ee7add3dad4725af6"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.544270 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fmq8p" event={"ID":"0d089cc1-fe1c-4ad4-a707-27c0e809649a","Type":"ContainerStarted","Data":"7a546a0c8b98e1540cb65f6782bfcaf02f884d8a9c4df7bb6e81685f7b3174d7"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.562124 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ks8v2" event={"ID":"8384bf14-e866-4e85-adaa-858bafca8741","Type":"ContainerStarted","Data":"be9e3aa5ee588bdadda8ed5a6d62a396a628af6b2518e04242eefa9021ea5aba"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.563086 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lcnnn" event={"ID":"fbf24c93-1112-4ca1-a16b-3025aa412eee","Type":"ContainerStarted","Data":"90c23b134643b83dac043d56be4de47f0c0dccfa12b7d34d1a237cc8edf8e35f"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.563935 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-rjzql" event={"ID":"f3b2043e-942e-4a01-95ae-cf8325faf27d","Type":"ContainerStarted","Data":"c4d53044be582810c081c55c6ea4bf8446ef8df1ec4b6bff1c8d81a11c7c8a5f"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.565302 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-cjsx7" event={"ID":"2c4fd4ad-d670-46d1-b1ea-71ec15ad5f8b","Type":"ContainerStarted","Data":"430fa7f334ef6fadd23f76ea68ce936822e984efd367958e4c3e2bac5d204b08"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.571428 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6n52l" event={"ID":"afe00a4b-526b-4a32-9d84-1321683d01ad","Type":"ContainerStarted","Data":"8784426b022545c828a0d6a06b5a753d719c037e9bc75252d799afa0ca29d931"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.573661 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" event={"ID":"0847bb67-0cb7-4616-b715-29b6f12f67b5","Type":"ContainerStarted","Data":"a255986c888b8d5cf49c452d7b4f3062861ed7cb28f3e965e92b1d5428262489"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.574512 4894 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-qq9k2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.574548 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" podUID="0847bb67-0cb7-4616-b715-29b6f12f67b5" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.575790 4894 generic.go:334] "Generic (PLEG): container finished" podID="39bada6f-31fa-46b6-bd31-6f43cc0f6aa4" containerID="ebe6fff805000df50420146e381e57dc67d28336b9c171fbb80a5c7dab1a3e76" exitCode=0 Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.576963 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" event={"ID":"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4","Type":"ContainerDied","Data":"ebe6fff805000df50420146e381e57dc67d28336b9c171fbb80a5c7dab1a3e76"} Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.579336 4894 patch_prober.go:28] interesting pod/console-operator-58897d9998-j5jl6 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.579374 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-j5jl6" podUID="b7b99437-26b8-4ac8-876e-60b2a92f70bb" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": dial tcp 10.217.0.15:8443: connect: connection refused" Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.594124 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:08 crc kubenswrapper[4894]: E1209 15:34:08.594545 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:09.094526276 +0000 UTC m=+143.413736955 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.594905 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:08 crc kubenswrapper[4894]: E1209 15:34:08.602414 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:09.102394558 +0000 UTC m=+143.421605297 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.696427 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:08 crc kubenswrapper[4894]: E1209 15:34:08.697812 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:09.197792927 +0000 UTC m=+143.517003586 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.726646 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r" podStartSLOduration=124.72661859 podStartE2EDuration="2m4.72661859s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:08.693881872 +0000 UTC m=+143.013092541" watchObservedRunningTime="2025-12-09 15:34:08.72661859 +0000 UTC m=+143.045829259" Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.798776 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:08 crc kubenswrapper[4894]: E1209 15:34:08.799081 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:09.299069624 +0000 UTC m=+143.618280293 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.827545 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" podStartSLOduration=124.827525017 podStartE2EDuration="2m4.827525017s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:08.72960152 +0000 UTC m=+143.048812189" watchObservedRunningTime="2025-12-09 15:34:08.827525017 +0000 UTC m=+143.146735686" Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.842190 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-fmq8p" podStartSLOduration=124.84217085 podStartE2EDuration="2m4.84217085s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:08.839437177 +0000 UTC m=+143.158647856" watchObservedRunningTime="2025-12-09 15:34:08.84217085 +0000 UTC m=+143.161381509" Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.860881 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-6n52l" podStartSLOduration=124.860863541 podStartE2EDuration="2m4.860863541s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:08.858653993 +0000 UTC m=+143.177864662" watchObservedRunningTime="2025-12-09 15:34:08.860863541 +0000 UTC m=+143.180074210" Dec 09 15:34:08 crc kubenswrapper[4894]: I1209 15:34:08.899513 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:08 crc kubenswrapper[4894]: E1209 15:34:08.899886 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:09.399871208 +0000 UTC m=+143.719081877 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.000604 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:09 crc kubenswrapper[4894]: E1209 15:34:09.000980 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:09.50096883 +0000 UTC m=+143.820179499 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.101629 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:09 crc kubenswrapper[4894]: E1209 15:34:09.102187 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:09.602168046 +0000 UTC m=+143.921378725 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.203007 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:09 crc kubenswrapper[4894]: E1209 15:34:09.203329 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:09.703316609 +0000 UTC m=+144.022527278 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:09 crc kubenswrapper[4894]: E1209 15:34:09.304520 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:09.804491463 +0000 UTC m=+144.123702162 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.304366 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.304879 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:09 crc kubenswrapper[4894]: E1209 15:34:09.305288 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:09.805273135 +0000 UTC m=+144.124483844 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.388466 4894 patch_prober.go:28] interesting pod/router-default-5444994796-sd4tb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 15:34:09 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 09 15:34:09 crc kubenswrapper[4894]: [+]process-running ok Dec 09 15:34:09 crc kubenswrapper[4894]: healthz check failed Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.388557 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sd4tb" podUID="a5600313-a000-42ce-94ca-898301d071fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.406746 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:09 crc kubenswrapper[4894]: E1209 15:34:09.406868 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:09.906830469 +0000 UTC m=+144.226041138 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.407085 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:09 crc kubenswrapper[4894]: E1209 15:34:09.407540 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:09.907529128 +0000 UTC m=+144.226739797 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.426750 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-ktwf9" Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.518296 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:09 crc kubenswrapper[4894]: E1209 15:34:09.518678 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:10.018663849 +0000 UTC m=+144.337874518 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.619522 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.640669 4894 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-pqs8f container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.25:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.648543 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f" podUID="1ec98242-7220-4547-b699-5e97bddfbca9" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.25:5443/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 15:34:09 crc kubenswrapper[4894]: E1209 15:34:09.640919 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:10.140908079 +0000 UTC m=+144.460118748 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.749928 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:09 crc kubenswrapper[4894]: E1209 15:34:09.750543 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:10.25052008 +0000 UTC m=+144.569730749 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.750706 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:09 crc kubenswrapper[4894]: E1209 15:34:09.751260 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:10.251251089 +0000 UTC m=+144.570461768 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.795785 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-d5smt" Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.825214 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-d5smt" podStartSLOduration=10.825194863 podStartE2EDuration="10.825194863s" podCreationTimestamp="2025-12-09 15:33:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:09.824967157 +0000 UTC m=+144.144177826" watchObservedRunningTime="2025-12-09 15:34:09.825194863 +0000 UTC m=+144.144405532" Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.841004 4894 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-56826 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.841043 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" podUID="171c863a-36f5-48c0-be81-752684537f0b" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.841342 4894 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-qq9k2 container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" start-of-body= Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.841369 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" podUID="0847bb67-0cb7-4616-b715-29b6f12f67b5" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.37:8080/healthz\": dial tcp 10.217.0.37:8080: connect: connection refused" Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.841921 4894 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-g79bw container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.17:6443/healthz\": dial tcp 10.217.0.17:6443: connect: connection refused" start-of-body= Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.841978 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" podUID="569bec4c-7397-468e-ad4f-dd69ee0ed023" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.17:6443/healthz\": dial tcp 10.217.0.17:6443: connect: connection refused" Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.853916 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:09 crc kubenswrapper[4894]: E1209 15:34:09.854389 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:10.354359266 +0000 UTC m=+144.673569935 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.855712 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:09 crc kubenswrapper[4894]: E1209 15:34:09.871854 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:10.371837204 +0000 UTC m=+144.691047873 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.885971 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ks8v2" podStartSLOduration=125.885951673 podStartE2EDuration="2m5.885951673s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:09.879517721 +0000 UTC m=+144.198728390" watchObservedRunningTime="2025-12-09 15:34:09.885951673 +0000 UTC m=+144.205162342" Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.977787 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:09 crc kubenswrapper[4894]: E1209 15:34:09.978606 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:10.478580039 +0000 UTC m=+144.797790718 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:09 crc kubenswrapper[4894]: I1209 15:34:09.996281 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-84l9b" podStartSLOduration=124.996262752 podStartE2EDuration="2m4.996262752s" podCreationTimestamp="2025-12-09 15:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:09.903493544 +0000 UTC m=+144.222704213" watchObservedRunningTime="2025-12-09 15:34:09.996262752 +0000 UTC m=+144.315473421" Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.081385 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:10 crc kubenswrapper[4894]: E1209 15:34:10.081794 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:10.581778667 +0000 UTC m=+144.900989336 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.149815 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-pqs8f" Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.182321 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:10 crc kubenswrapper[4894]: E1209 15:34:10.182502 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:10.682477408 +0000 UTC m=+145.001688077 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.182599 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:10 crc kubenswrapper[4894]: E1209 15:34:10.183079 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:10.683065085 +0000 UTC m=+145.002275764 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.210121 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-5px2s" podStartSLOduration=125.21010393 podStartE2EDuration="2m5.21010393s" podCreationTimestamp="2025-12-09 15:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:09.996394836 +0000 UTC m=+144.315605495" watchObservedRunningTime="2025-12-09 15:34:10.21010393 +0000 UTC m=+144.529314599" Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.283673 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:10 crc kubenswrapper[4894]: E1209 15:34:10.284385 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:10.784364242 +0000 UTC m=+145.103574911 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.385754 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:10 crc kubenswrapper[4894]: E1209 15:34:10.386190 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:10.886172143 +0000 UTC m=+145.205382882 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.389402 4894 patch_prober.go:28] interesting pod/router-default-5444994796-sd4tb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 15:34:10 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 09 15:34:10 crc kubenswrapper[4894]: [+]process-running ok Dec 09 15:34:10 crc kubenswrapper[4894]: healthz check failed Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.389463 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sd4tb" podUID="a5600313-a000-42ce-94ca-898301d071fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.486359 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:10 crc kubenswrapper[4894]: E1209 15:34:10.486490 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:10.986472444 +0000 UTC m=+145.305683103 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.486726 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:10 crc kubenswrapper[4894]: E1209 15:34:10.487018 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:10.987010849 +0000 UTC m=+145.306221518 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.588104 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:10 crc kubenswrapper[4894]: E1209 15:34:10.588330 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:11.088304486 +0000 UTC m=+145.407515155 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.588383 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:10 crc kubenswrapper[4894]: E1209 15:34:10.589129 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:11.089112538 +0000 UTC m=+145.408323207 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.689956 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:10 crc kubenswrapper[4894]: E1209 15:34:10.690065 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:11.190047586 +0000 UTC m=+145.509258255 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.690262 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:10 crc kubenswrapper[4894]: E1209 15:34:10.690523 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:11.190514518 +0000 UTC m=+145.509725187 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.791743 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:10 crc kubenswrapper[4894]: E1209 15:34:10.792122 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:11.292105764 +0000 UTC m=+145.611316433 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.842187 4894 patch_prober.go:28] interesting pod/console-operator-58897d9998-j5jl6 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.842248 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-j5jl6" podUID="b7b99437-26b8-4ac8-876e-60b2a92f70bb" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.15:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.874052 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" event={"ID":"1d071a99-3aa2-4109-a037-d05e278d1206","Type":"ContainerStarted","Data":"1db0df59ce436c31b8b9d14c3f2d01cfec7c4e0324ec523d2e3e6fa8b7f53bff"} Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.874097 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" event={"ID":"1d071a99-3aa2-4109-a037-d05e278d1206","Type":"ContainerStarted","Data":"1f79ee70f29c52a0d3ad5d35b46bfae342bf455433a057cdd0100decb70ab842"} Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.892790 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:10 crc kubenswrapper[4894]: E1209 15:34:10.893175 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:11.393161745 +0000 UTC m=+145.712372414 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.898971 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m8gtv" event={"ID":"7940af9e-4cd3-4def-bdbe-c8a6ac61bb4d","Type":"ContainerStarted","Data":"34646f2b7b30e3e98d9d17cec4b52e7c7458ad0bef327fa6c30b6edbfb80f8fc"} Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.899783 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m8gtv" Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.911923 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lcnnn" event={"ID":"fbf24c93-1112-4ca1-a16b-3025aa412eee","Type":"ContainerStarted","Data":"7fbf2cf6626da75d4b8008246d0f5f9c20f6d158f9f516839f7b8e80ad785de7"} Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.916852 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" event={"ID":"f3b629b0-15d9-4e1d-908c-b21f07444427","Type":"ContainerStarted","Data":"ea6d52ee21481329ab934b2985b808be4f52f04960d6c20e4e0462026850a6a7"} Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.918723 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-d5smt" event={"ID":"20c26380-64dd-4fe1-a9d5-febed69bea4e","Type":"ContainerStarted","Data":"a0e5b3e5b8fd24401963d5b98fb36685ca49a21eb8a7983c6efd6479a34cdf46"} Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.920646 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" event={"ID":"66f45f17-addc-4737-a47c-1ab34a289659","Type":"ContainerStarted","Data":"b5f3a1555e79909f3764685d070fd0e3453425d5d07c7ea0f0a09905209ac548"} Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.933377 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-7nlp9" event={"ID":"a1be956b-214a-400f-8f5b-201fbe268090","Type":"ContainerStarted","Data":"4fd716fde99e3608b7048de4eada47fb110be0565976e8fc25a59c5906fe41ab"} Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.938405 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" event={"ID":"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4","Type":"ContainerStarted","Data":"2e5e5a359876fc65149528131cec543929d5ad6116f87c1de6bae4c65222cca5"} Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.938453 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" event={"ID":"39bada6f-31fa-46b6-bd31-6f43cc0f6aa4","Type":"ContainerStarted","Data":"44ac6aed33b0e17bb69f12a7b242b16c50fb1efa98ae0062d30fdd8613797046"} Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.938803 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m8gtv" podStartSLOduration=125.938787999 podStartE2EDuration="2m5.938787999s" podCreationTimestamp="2025-12-09 15:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:10.935774808 +0000 UTC m=+145.254985507" watchObservedRunningTime="2025-12-09 15:34:10.938787999 +0000 UTC m=+145.257998678" Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.971385 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-7nlp9" podStartSLOduration=125.971367473 podStartE2EDuration="2m5.971367473s" podCreationTimestamp="2025-12-09 15:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:10.970101789 +0000 UTC m=+145.289312458" watchObservedRunningTime="2025-12-09 15:34:10.971367473 +0000 UTC m=+145.290578142" Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.977816 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.993414 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" podStartSLOduration=125.993396974 podStartE2EDuration="2m5.993396974s" podCreationTimestamp="2025-12-09 15:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:10.991916364 +0000 UTC m=+145.311127043" watchObservedRunningTime="2025-12-09 15:34:10.993396974 +0000 UTC m=+145.312607643" Dec 09 15:34:10 crc kubenswrapper[4894]: I1209 15:34:10.994182 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:10 crc kubenswrapper[4894]: E1209 15:34:10.995795 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:11.495779978 +0000 UTC m=+145.814990647 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.025683 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lcnnn" podStartSLOduration=126.025670009 podStartE2EDuration="2m6.025670009s" podCreationTimestamp="2025-12-09 15:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:11.024271832 +0000 UTC m=+145.343482501" watchObservedRunningTime="2025-12-09 15:34:11.025670009 +0000 UTC m=+145.344880678" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.066048 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-rvzr9" podStartSLOduration=126.066031353 podStartE2EDuration="2m6.066031353s" podCreationTimestamp="2025-12-09 15:32:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:11.064546343 +0000 UTC m=+145.383757012" watchObservedRunningTime="2025-12-09 15:34:11.066031353 +0000 UTC m=+145.385242022" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.097054 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:11 crc kubenswrapper[4894]: E1209 15:34:11.097412 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:11.597400325 +0000 UTC m=+145.916610994 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.137816 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" podStartSLOduration=127.137801159 podStartE2EDuration="2m7.137801159s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:11.136396671 +0000 UTC m=+145.455607340" watchObservedRunningTime="2025-12-09 15:34:11.137801159 +0000 UTC m=+145.457011828" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.198258 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:11 crc kubenswrapper[4894]: E1209 15:34:11.198594 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:11.698574219 +0000 UTC m=+146.017784888 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.300377 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:11 crc kubenswrapper[4894]: E1209 15:34:11.301982 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:11.801964413 +0000 UTC m=+146.121175082 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.309596 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hq8h6"] Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.310815 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hq8h6" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.320009 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.329374 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hq8h6"] Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.395787 4894 patch_prober.go:28] interesting pod/router-default-5444994796-sd4tb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 15:34:11 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 09 15:34:11 crc kubenswrapper[4894]: [+]process-running ok Dec 09 15:34:11 crc kubenswrapper[4894]: healthz check failed Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.395847 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sd4tb" podUID="a5600313-a000-42ce-94ca-898301d071fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.405154 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.405477 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4k9sg\" (UniqueName: \"kubernetes.io/projected/4397f08d-278a-4e2d-9664-e44161c96638-kube-api-access-4k9sg\") pod \"certified-operators-hq8h6\" (UID: \"4397f08d-278a-4e2d-9664-e44161c96638\") " pod="openshift-marketplace/certified-operators-hq8h6" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.405553 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4397f08d-278a-4e2d-9664-e44161c96638-catalog-content\") pod \"certified-operators-hq8h6\" (UID: \"4397f08d-278a-4e2d-9664-e44161c96638\") " pod="openshift-marketplace/certified-operators-hq8h6" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.405582 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4397f08d-278a-4e2d-9664-e44161c96638-utilities\") pod \"certified-operators-hq8h6\" (UID: \"4397f08d-278a-4e2d-9664-e44161c96638\") " pod="openshift-marketplace/certified-operators-hq8h6" Dec 09 15:34:11 crc kubenswrapper[4894]: E1209 15:34:11.405755 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:11.905732307 +0000 UTC m=+146.224942976 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.503312 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-ff286"] Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.504339 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ff286" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.507015 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4397f08d-278a-4e2d-9664-e44161c96638-catalog-content\") pod \"certified-operators-hq8h6\" (UID: \"4397f08d-278a-4e2d-9664-e44161c96638\") " pod="openshift-marketplace/certified-operators-hq8h6" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.507047 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4397f08d-278a-4e2d-9664-e44161c96638-utilities\") pod \"certified-operators-hq8h6\" (UID: \"4397f08d-278a-4e2d-9664-e44161c96638\") " pod="openshift-marketplace/certified-operators-hq8h6" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.507126 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.507145 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4k9sg\" (UniqueName: \"kubernetes.io/projected/4397f08d-278a-4e2d-9664-e44161c96638-kube-api-access-4k9sg\") pod \"certified-operators-hq8h6\" (UID: \"4397f08d-278a-4e2d-9664-e44161c96638\") " pod="openshift-marketplace/certified-operators-hq8h6" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.508623 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4397f08d-278a-4e2d-9664-e44161c96638-catalog-content\") pod \"certified-operators-hq8h6\" (UID: \"4397f08d-278a-4e2d-9664-e44161c96638\") " pod="openshift-marketplace/certified-operators-hq8h6" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.508889 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4397f08d-278a-4e2d-9664-e44161c96638-utilities\") pod \"certified-operators-hq8h6\" (UID: \"4397f08d-278a-4e2d-9664-e44161c96638\") " pod="openshift-marketplace/certified-operators-hq8h6" Dec 09 15:34:11 crc kubenswrapper[4894]: E1209 15:34:11.509153 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:12.009140721 +0000 UTC m=+146.328351390 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:11 crc kubenswrapper[4894]: W1209 15:34:11.509448 4894 reflector.go:561] object-"openshift-marketplace"/"community-operators-dockercfg-dmngl": failed to list *v1.Secret: secrets "community-operators-dockercfg-dmngl" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-marketplace": no relationship found between node 'crc' and this object Dec 09 15:34:11 crc kubenswrapper[4894]: E1209 15:34:11.509472 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-marketplace\"/\"community-operators-dockercfg-dmngl\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"community-operators-dockercfg-dmngl\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-marketplace\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.530716 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ff286"] Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.568120 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4k9sg\" (UniqueName: \"kubernetes.io/projected/4397f08d-278a-4e2d-9664-e44161c96638-kube-api-access-4k9sg\") pod \"certified-operators-hq8h6\" (UID: \"4397f08d-278a-4e2d-9664-e44161c96638\") " pod="openshift-marketplace/certified-operators-hq8h6" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.608712 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.609024 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c278157d-d806-4fa1-a597-076d95a286e2-catalog-content\") pod \"community-operators-ff286\" (UID: \"c278157d-d806-4fa1-a597-076d95a286e2\") " pod="openshift-marketplace/community-operators-ff286" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.609065 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4d2d\" (UniqueName: \"kubernetes.io/projected/c278157d-d806-4fa1-a597-076d95a286e2-kube-api-access-r4d2d\") pod \"community-operators-ff286\" (UID: \"c278157d-d806-4fa1-a597-076d95a286e2\") " pod="openshift-marketplace/community-operators-ff286" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.609102 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c278157d-d806-4fa1-a597-076d95a286e2-utilities\") pod \"community-operators-ff286\" (UID: \"c278157d-d806-4fa1-a597-076d95a286e2\") " pod="openshift-marketplace/community-operators-ff286" Dec 09 15:34:11 crc kubenswrapper[4894]: E1209 15:34:11.609282 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:12.109263997 +0000 UTC m=+146.428474746 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.615836 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.643981 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hq8h6" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.679912 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.680305 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.681706 4894 patch_prober.go:28] interesting pod/console-f9d7485db-crhvg container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.681856 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-crhvg" podUID="9d548d70-c4af-4333-8cea-f6bef5722ce5" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.685158 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-fxq7l"] Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.686432 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fxq7l" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.711611 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqpnd\" (UniqueName: \"kubernetes.io/projected/e042f2bf-1069-42a4-aef2-72ec5da57c99-kube-api-access-gqpnd\") pod \"certified-operators-fxq7l\" (UID: \"e042f2bf-1069-42a4-aef2-72ec5da57c99\") " pod="openshift-marketplace/certified-operators-fxq7l" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.711672 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e042f2bf-1069-42a4-aef2-72ec5da57c99-utilities\") pod \"certified-operators-fxq7l\" (UID: \"e042f2bf-1069-42a4-aef2-72ec5da57c99\") " pod="openshift-marketplace/certified-operators-fxq7l" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.711699 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c278157d-d806-4fa1-a597-076d95a286e2-catalog-content\") pod \"community-operators-ff286\" (UID: \"c278157d-d806-4fa1-a597-076d95a286e2\") " pod="openshift-marketplace/community-operators-ff286" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.711717 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4d2d\" (UniqueName: \"kubernetes.io/projected/c278157d-d806-4fa1-a597-076d95a286e2-kube-api-access-r4d2d\") pod \"community-operators-ff286\" (UID: \"c278157d-d806-4fa1-a597-076d95a286e2\") " pod="openshift-marketplace/community-operators-ff286" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.711754 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c278157d-d806-4fa1-a597-076d95a286e2-utilities\") pod \"community-operators-ff286\" (UID: \"c278157d-d806-4fa1-a597-076d95a286e2\") " pod="openshift-marketplace/community-operators-ff286" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.711845 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.711867 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e042f2bf-1069-42a4-aef2-72ec5da57c99-catalog-content\") pod \"certified-operators-fxq7l\" (UID: \"e042f2bf-1069-42a4-aef2-72ec5da57c99\") " pod="openshift-marketplace/certified-operators-fxq7l" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.713593 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c278157d-d806-4fa1-a597-076d95a286e2-catalog-content\") pod \"community-operators-ff286\" (UID: \"c278157d-d806-4fa1-a597-076d95a286e2\") " pod="openshift-marketplace/community-operators-ff286" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.714391 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c278157d-d806-4fa1-a597-076d95a286e2-utilities\") pod \"community-operators-ff286\" (UID: \"c278157d-d806-4fa1-a597-076d95a286e2\") " pod="openshift-marketplace/community-operators-ff286" Dec 09 15:34:11 crc kubenswrapper[4894]: E1209 15:34:11.715056 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:12.215036225 +0000 UTC m=+146.534246964 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.762013 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4d2d\" (UniqueName: \"kubernetes.io/projected/c278157d-d806-4fa1-a597-076d95a286e2-kube-api-access-r4d2d\") pod \"community-operators-ff286\" (UID: \"c278157d-d806-4fa1-a597-076d95a286e2\") " pod="openshift-marketplace/community-operators-ff286" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.813374 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.813801 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqpnd\" (UniqueName: \"kubernetes.io/projected/e042f2bf-1069-42a4-aef2-72ec5da57c99-kube-api-access-gqpnd\") pod \"certified-operators-fxq7l\" (UID: \"e042f2bf-1069-42a4-aef2-72ec5da57c99\") " pod="openshift-marketplace/certified-operators-fxq7l" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.813898 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e042f2bf-1069-42a4-aef2-72ec5da57c99-utilities\") pod \"certified-operators-fxq7l\" (UID: \"e042f2bf-1069-42a4-aef2-72ec5da57c99\") " pod="openshift-marketplace/certified-operators-fxq7l" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.814062 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e042f2bf-1069-42a4-aef2-72ec5da57c99-catalog-content\") pod \"certified-operators-fxq7l\" (UID: \"e042f2bf-1069-42a4-aef2-72ec5da57c99\") " pod="openshift-marketplace/certified-operators-fxq7l" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.814621 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e042f2bf-1069-42a4-aef2-72ec5da57c99-catalog-content\") pod \"certified-operators-fxq7l\" (UID: \"e042f2bf-1069-42a4-aef2-72ec5da57c99\") " pod="openshift-marketplace/certified-operators-fxq7l" Dec 09 15:34:11 crc kubenswrapper[4894]: E1209 15:34:11.814796 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:12.31478158 +0000 UTC m=+146.633992249 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.815232 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e042f2bf-1069-42a4-aef2-72ec5da57c99-utilities\") pod \"certified-operators-fxq7l\" (UID: \"e042f2bf-1069-42a4-aef2-72ec5da57c99\") " pod="openshift-marketplace/certified-operators-fxq7l" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.819824 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fxq7l"] Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.855161 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqpnd\" (UniqueName: \"kubernetes.io/projected/e042f2bf-1069-42a4-aef2-72ec5da57c99-kube-api-access-gqpnd\") pod \"certified-operators-fxq7l\" (UID: \"e042f2bf-1069-42a4-aef2-72ec5da57c99\") " pod="openshift-marketplace/certified-operators-fxq7l" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.900606 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-s4rbb"] Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.901518 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s4rbb" Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.918226 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:11 crc kubenswrapper[4894]: E1209 15:34:11.918501 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:12.418490253 +0000 UTC m=+146.737700922 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:11 crc kubenswrapper[4894]: I1209 15:34:11.920434 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s4rbb"] Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.000804 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" event={"ID":"1d071a99-3aa2-4109-a037-d05e278d1206","Type":"ContainerStarted","Data":"c17e3ff4e321f7e1ac47d789352e51d36cdc0482e7e3fa59bae6efd8ef5f7344"} Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.019115 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.019339 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d076e7-ca39-4337-bacc-ea3984f3bb5d-utilities\") pod \"community-operators-s4rbb\" (UID: \"46d076e7-ca39-4337-bacc-ea3984f3bb5d\") " pod="openshift-marketplace/community-operators-s4rbb" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.019383 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-trr6z\" (UniqueName: \"kubernetes.io/projected/46d076e7-ca39-4337-bacc-ea3984f3bb5d-kube-api-access-trr6z\") pod \"community-operators-s4rbb\" (UID: \"46d076e7-ca39-4337-bacc-ea3984f3bb5d\") " pod="openshift-marketplace/community-operators-s4rbb" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.019447 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d076e7-ca39-4337-bacc-ea3984f3bb5d-catalog-content\") pod \"community-operators-s4rbb\" (UID: \"46d076e7-ca39-4337-bacc-ea3984f3bb5d\") " pod="openshift-marketplace/community-operators-s4rbb" Dec 09 15:34:12 crc kubenswrapper[4894]: E1209 15:34:12.019597 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:12.519556944 +0000 UTC m=+146.838767613 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.031711 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fxq7l" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.085305 4894 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.120950 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.120987 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d076e7-ca39-4337-bacc-ea3984f3bb5d-catalog-content\") pod \"community-operators-s4rbb\" (UID: \"46d076e7-ca39-4337-bacc-ea3984f3bb5d\") " pod="openshift-marketplace/community-operators-s4rbb" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.121212 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d076e7-ca39-4337-bacc-ea3984f3bb5d-utilities\") pod \"community-operators-s4rbb\" (UID: \"46d076e7-ca39-4337-bacc-ea3984f3bb5d\") " pod="openshift-marketplace/community-operators-s4rbb" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.121305 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-trr6z\" (UniqueName: \"kubernetes.io/projected/46d076e7-ca39-4337-bacc-ea3984f3bb5d-kube-api-access-trr6z\") pod \"community-operators-s4rbb\" (UID: \"46d076e7-ca39-4337-bacc-ea3984f3bb5d\") " pod="openshift-marketplace/community-operators-s4rbb" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.122787 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d076e7-ca39-4337-bacc-ea3984f3bb5d-utilities\") pod \"community-operators-s4rbb\" (UID: \"46d076e7-ca39-4337-bacc-ea3984f3bb5d\") " pod="openshift-marketplace/community-operators-s4rbb" Dec 09 15:34:12 crc kubenswrapper[4894]: E1209 15:34:12.123281 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-12-09 15:34:12.623265056 +0000 UTC m=+146.942475715 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-k8jmv" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.123802 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d076e7-ca39-4337-bacc-ea3984f3bb5d-catalog-content\") pod \"community-operators-s4rbb\" (UID: \"46d076e7-ca39-4337-bacc-ea3984f3bb5d\") " pod="openshift-marketplace/community-operators-s4rbb" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.151291 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-trr6z\" (UniqueName: \"kubernetes.io/projected/46d076e7-ca39-4337-bacc-ea3984f3bb5d-kube-api-access-trr6z\") pod \"community-operators-s4rbb\" (UID: \"46d076e7-ca39-4337-bacc-ea3984f3bb5d\") " pod="openshift-marketplace/community-operators-s4rbb" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.227974 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:12 crc kubenswrapper[4894]: E1209 15:34:12.228255 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-12-09 15:34:12.728241833 +0000 UTC m=+147.047452502 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.236206 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ks8v2" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.261116 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.261171 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.273717 4894 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-12-09T15:34:12.085522774Z","Handler":null,"Name":""} Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.280989 4894 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.281026 4894 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.305873 4894 patch_prober.go:28] interesting pod/downloads-7954f5f757-vtbc8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.305945 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vtbc8" podUID="caed7380-e514-4f40-ae98-0be4a2fe61c3" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.305959 4894 patch_prober.go:28] interesting pod/downloads-7954f5f757-vtbc8 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.306005 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-vtbc8" podUID="caed7380-e514-4f40-ae98-0be4a2fe61c3" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.331465 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.335317 4894 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.335355 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.356690 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-k8jmv\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.383985 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hq8h6"] Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.388369 4894 patch_prober.go:28] interesting pod/router-default-5444994796-sd4tb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 15:34:12 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 09 15:34:12 crc kubenswrapper[4894]: [+]process-running ok Dec 09 15:34:12 crc kubenswrapper[4894]: healthz check failed Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.388432 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sd4tb" podUID="a5600313-a000-42ce-94ca-898301d071fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.433211 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.493012 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.493758 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.493833 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ff286" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.494264 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s4rbb" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.529989 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.547771 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ks8v2" Dec 09 15:34:12 crc kubenswrapper[4894]: I1209 15:34:12.579863 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-fxq7l"] Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.027907 4894 generic.go:334] "Generic (PLEG): container finished" podID="f67460f9-bc45-47e0-b186-f66dea9a578c" containerID="d41ad5679ea1a20ef7453e9a31ff71dd7d5f8395803d542aa6e12f46dafa93f3" exitCode=0 Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.028104 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r" event={"ID":"f67460f9-bc45-47e0-b186-f66dea9a578c","Type":"ContainerDied","Data":"d41ad5679ea1a20ef7453e9a31ff71dd7d5f8395803d542aa6e12f46dafa93f3"} Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.042328 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.042374 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.047751 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.053281 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxq7l" event={"ID":"e042f2bf-1069-42a4-aef2-72ec5da57c99","Type":"ContainerStarted","Data":"f338cf6ba9baffe040fc08bc935fda57c2613b4df4d52db7f30c58c48997057f"} Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.068025 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-s4rbb"] Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.069080 4894 generic.go:334] "Generic (PLEG): container finished" podID="4397f08d-278a-4e2d-9664-e44161c96638" containerID="cfba892241222e18b3fdce72ef9bb2cefd0c77fbcb02ba2d7ad49f5d06c56b48" exitCode=0 Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.069216 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hq8h6" event={"ID":"4397f08d-278a-4e2d-9664-e44161c96638","Type":"ContainerDied","Data":"cfba892241222e18b3fdce72ef9bb2cefd0c77fbcb02ba2d7ad49f5d06c56b48"} Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.069302 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hq8h6" event={"ID":"4397f08d-278a-4e2d-9664-e44161c96638","Type":"ContainerStarted","Data":"00e10a54574bc8485c5bef256c44e50c8eb056dafe9840669ae7df957ec49776"} Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.072287 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.074484 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.090862 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" event={"ID":"1d071a99-3aa2-4109-a037-d05e278d1206","Type":"ContainerStarted","Data":"db41c40d320a5d845ad1994858d8514a595018c28d001d98d83d50347543d5b6"} Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.126287 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k8jmv"] Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.133671 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.151717 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.151826 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.162279 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.191833 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-l6cb5" podStartSLOduration=13.191805645 podStartE2EDuration="13.191805645s" podCreationTimestamp="2025-12-09 15:34:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:13.16593555 +0000 UTC m=+147.485146239" watchObservedRunningTime="2025-12-09 15:34:13.191805645 +0000 UTC m=+147.511016324" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.193911 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.241957 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-j5jl6" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.305455 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-nxlqt"] Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.306664 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nxlqt" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.309876 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.321209 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.327974 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.343768 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nxlqt"] Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.379116 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-ff286"] Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.386510 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-sd4tb" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.389223 4894 patch_prober.go:28] interesting pod/router-default-5444994796-sd4tb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 15:34:13 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 09 15:34:13 crc kubenswrapper[4894]: [+]process-running ok Dec 09 15:34:13 crc kubenswrapper[4894]: healthz check failed Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.389262 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sd4tb" podUID="a5600313-a000-42ce-94ca-898301d071fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.454791 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-whpnq\" (UniqueName: \"kubernetes.io/projected/2421584b-8a4b-4cb1-97c0-cb41d8ec16b2-kube-api-access-whpnq\") pod \"redhat-marketplace-nxlqt\" (UID: \"2421584b-8a4b-4cb1-97c0-cb41d8ec16b2\") " pod="openshift-marketplace/redhat-marketplace-nxlqt" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.454828 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2421584b-8a4b-4cb1-97c0-cb41d8ec16b2-utilities\") pod \"redhat-marketplace-nxlqt\" (UID: \"2421584b-8a4b-4cb1-97c0-cb41d8ec16b2\") " pod="openshift-marketplace/redhat-marketplace-nxlqt" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.454896 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2421584b-8a4b-4cb1-97c0-cb41d8ec16b2-catalog-content\") pod \"redhat-marketplace-nxlqt\" (UID: \"2421584b-8a4b-4cb1-97c0-cb41d8ec16b2\") " pod="openshift-marketplace/redhat-marketplace-nxlqt" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.556174 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-whpnq\" (UniqueName: \"kubernetes.io/projected/2421584b-8a4b-4cb1-97c0-cb41d8ec16b2-kube-api-access-whpnq\") pod \"redhat-marketplace-nxlqt\" (UID: \"2421584b-8a4b-4cb1-97c0-cb41d8ec16b2\") " pod="openshift-marketplace/redhat-marketplace-nxlqt" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.556230 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2421584b-8a4b-4cb1-97c0-cb41d8ec16b2-utilities\") pod \"redhat-marketplace-nxlqt\" (UID: \"2421584b-8a4b-4cb1-97c0-cb41d8ec16b2\") " pod="openshift-marketplace/redhat-marketplace-nxlqt" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.556308 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2421584b-8a4b-4cb1-97c0-cb41d8ec16b2-catalog-content\") pod \"redhat-marketplace-nxlqt\" (UID: \"2421584b-8a4b-4cb1-97c0-cb41d8ec16b2\") " pod="openshift-marketplace/redhat-marketplace-nxlqt" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.557545 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2421584b-8a4b-4cb1-97c0-cb41d8ec16b2-utilities\") pod \"redhat-marketplace-nxlqt\" (UID: \"2421584b-8a4b-4cb1-97c0-cb41d8ec16b2\") " pod="openshift-marketplace/redhat-marketplace-nxlqt" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.557847 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2421584b-8a4b-4cb1-97c0-cb41d8ec16b2-catalog-content\") pod \"redhat-marketplace-nxlqt\" (UID: \"2421584b-8a4b-4cb1-97c0-cb41d8ec16b2\") " pod="openshift-marketplace/redhat-marketplace-nxlqt" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.599555 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-whpnq\" (UniqueName: \"kubernetes.io/projected/2421584b-8a4b-4cb1-97c0-cb41d8ec16b2-kube-api-access-whpnq\") pod \"redhat-marketplace-nxlqt\" (UID: \"2421584b-8a4b-4cb1-97c0-cb41d8ec16b2\") " pod="openshift-marketplace/redhat-marketplace-nxlqt" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.635164 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nxlqt" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.697720 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q58q9"] Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.698950 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q58q9" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.726190 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q58q9"] Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.735459 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.773166 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.773968 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.777077 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.782536 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.783000 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.859792 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee374eee-998a-48f2-acdf-39daa5ce2fdd-utilities\") pod \"redhat-marketplace-q58q9\" (UID: \"ee374eee-998a-48f2-acdf-39daa5ce2fdd\") " pod="openshift-marketplace/redhat-marketplace-q58q9" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.859846 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee374eee-998a-48f2-acdf-39daa5ce2fdd-catalog-content\") pod \"redhat-marketplace-q58q9\" (UID: \"ee374eee-998a-48f2-acdf-39daa5ce2fdd\") " pod="openshift-marketplace/redhat-marketplace-q58q9" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.859933 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwmhn\" (UniqueName: \"kubernetes.io/projected/ee374eee-998a-48f2-acdf-39daa5ce2fdd-kube-api-access-hwmhn\") pod \"redhat-marketplace-q58q9\" (UID: \"ee374eee-998a-48f2-acdf-39daa5ce2fdd\") " pod="openshift-marketplace/redhat-marketplace-q58q9" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.889095 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.889546 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.902417 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:13 crc kubenswrapper[4894]: W1209 15:34:13.907085 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-e65d258fe1847920e94d01328f86d29c6299eedd0c77de82a47e9026f3762aee WatchSource:0}: Error finding container e65d258fe1847920e94d01328f86d29c6299eedd0c77de82a47e9026f3762aee: Status 404 returned error can't find the container with id e65d258fe1847920e94d01328f86d29c6299eedd0c77de82a47e9026f3762aee Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.960619 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwmhn\" (UniqueName: \"kubernetes.io/projected/ee374eee-998a-48f2-acdf-39daa5ce2fdd-kube-api-access-hwmhn\") pod \"redhat-marketplace-q58q9\" (UID: \"ee374eee-998a-48f2-acdf-39daa5ce2fdd\") " pod="openshift-marketplace/redhat-marketplace-q58q9" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.960724 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee374eee-998a-48f2-acdf-39daa5ce2fdd-utilities\") pod \"redhat-marketplace-q58q9\" (UID: \"ee374eee-998a-48f2-acdf-39daa5ce2fdd\") " pod="openshift-marketplace/redhat-marketplace-q58q9" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.960747 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee374eee-998a-48f2-acdf-39daa5ce2fdd-catalog-content\") pod \"redhat-marketplace-q58q9\" (UID: \"ee374eee-998a-48f2-acdf-39daa5ce2fdd\") " pod="openshift-marketplace/redhat-marketplace-q58q9" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.960772 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5832b82d-0bc8-434a-9907-7554ca34a3a6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5832b82d-0bc8-434a-9907-7554ca34a3a6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.960793 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5832b82d-0bc8-434a-9907-7554ca34a3a6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5832b82d-0bc8-434a-9907-7554ca34a3a6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.961159 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee374eee-998a-48f2-acdf-39daa5ce2fdd-utilities\") pod \"redhat-marketplace-q58q9\" (UID: \"ee374eee-998a-48f2-acdf-39daa5ce2fdd\") " pod="openshift-marketplace/redhat-marketplace-q58q9" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.962070 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee374eee-998a-48f2-acdf-39daa5ce2fdd-catalog-content\") pod \"redhat-marketplace-q58q9\" (UID: \"ee374eee-998a-48f2-acdf-39daa5ce2fdd\") " pod="openshift-marketplace/redhat-marketplace-q58q9" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.981698 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwmhn\" (UniqueName: \"kubernetes.io/projected/ee374eee-998a-48f2-acdf-39daa5ce2fdd-kube-api-access-hwmhn\") pod \"redhat-marketplace-q58q9\" (UID: \"ee374eee-998a-48f2-acdf-39daa5ce2fdd\") " pod="openshift-marketplace/redhat-marketplace-q58q9" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.996658 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:13 crc kubenswrapper[4894]: I1209 15:34:13.997756 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.018384 4894 patch_prober.go:28] interesting pod/apiserver-76f77b778f-bcfbr container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 09 15:34:14 crc kubenswrapper[4894]: [+]log ok Dec 09 15:34:14 crc kubenswrapper[4894]: [+]etcd ok Dec 09 15:34:14 crc kubenswrapper[4894]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 09 15:34:14 crc kubenswrapper[4894]: [+]poststarthook/generic-apiserver-start-informers ok Dec 09 15:34:14 crc kubenswrapper[4894]: [+]poststarthook/max-in-flight-filter ok Dec 09 15:34:14 crc kubenswrapper[4894]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 09 15:34:14 crc kubenswrapper[4894]: [+]poststarthook/image.openshift.io-apiserver-caches ok Dec 09 15:34:14 crc kubenswrapper[4894]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Dec 09 15:34:14 crc kubenswrapper[4894]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Dec 09 15:34:14 crc kubenswrapper[4894]: [+]poststarthook/project.openshift.io-projectcache ok Dec 09 15:34:14 crc kubenswrapper[4894]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Dec 09 15:34:14 crc kubenswrapper[4894]: [+]poststarthook/openshift.io-startinformers ok Dec 09 15:34:14 crc kubenswrapper[4894]: [+]poststarthook/openshift.io-restmapperupdater ok Dec 09 15:34:14 crc kubenswrapper[4894]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 09 15:34:14 crc kubenswrapper[4894]: livez check failed Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.018435 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" podUID="39bada6f-31fa-46b6-bd31-6f43cc0f6aa4" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.032664 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q58q9" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.036804 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-nxlqt"] Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.061673 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5832b82d-0bc8-434a-9907-7554ca34a3a6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5832b82d-0bc8-434a-9907-7554ca34a3a6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.061722 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5832b82d-0bc8-434a-9907-7554ca34a3a6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5832b82d-0bc8-434a-9907-7554ca34a3a6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.061829 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5832b82d-0bc8-434a-9907-7554ca34a3a6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"5832b82d-0bc8-434a-9907-7554ca34a3a6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.083447 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5832b82d-0bc8-434a-9907-7554ca34a3a6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"5832b82d-0bc8-434a-9907-7554ca34a3a6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.093039 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.101610 4894 generic.go:334] "Generic (PLEG): container finished" podID="c278157d-d806-4fa1-a597-076d95a286e2" containerID="7579df81a9780e5a1398c81e9f7ae75cb015cb52640413bdfb24749a8ea432a0" exitCode=0 Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.101680 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ff286" event={"ID":"c278157d-d806-4fa1-a597-076d95a286e2","Type":"ContainerDied","Data":"7579df81a9780e5a1398c81e9f7ae75cb015cb52640413bdfb24749a8ea432a0"} Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.101725 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ff286" event={"ID":"c278157d-d806-4fa1-a597-076d95a286e2","Type":"ContainerStarted","Data":"8240059115ee9c7c5aa57f051668183b7740875bf4567a338cc19b74cce4febc"} Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.104068 4894 generic.go:334] "Generic (PLEG): container finished" podID="e042f2bf-1069-42a4-aef2-72ec5da57c99" containerID="5cb2340301f99f23c4d77ce927ea630a9c59e1adb619f5e4605d2d3f5a8a65cf" exitCode=0 Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.104375 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxq7l" event={"ID":"e042f2bf-1069-42a4-aef2-72ec5da57c99","Type":"ContainerDied","Data":"5cb2340301f99f23c4d77ce927ea630a9c59e1adb619f5e4605d2d3f5a8a65cf"} Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.109575 4894 generic.go:334] "Generic (PLEG): container finished" podID="46d076e7-ca39-4337-bacc-ea3984f3bb5d" containerID="814bfae2a1ab16cee44409ed752fbecd54395276985b4cda425a0567223ddff5" exitCode=0 Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.133765 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.134513 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s4rbb" event={"ID":"46d076e7-ca39-4337-bacc-ea3984f3bb5d","Type":"ContainerDied","Data":"814bfae2a1ab16cee44409ed752fbecd54395276985b4cda425a0567223ddff5"} Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.134543 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s4rbb" event={"ID":"46d076e7-ca39-4337-bacc-ea3984f3bb5d","Type":"ContainerStarted","Data":"dd15011d9c6afe0129e5e4b73e8a3501887648369064fb7f075281ca284a2c08"} Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.134555 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxlqt" event={"ID":"2421584b-8a4b-4cb1-97c0-cb41d8ec16b2","Type":"ContainerStarted","Data":"712254f1d9317e269ef0c61a6c33e1f85a6bb1273c6989b297bd36a8e9691081"} Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.134566 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"e65d258fe1847920e94d01328f86d29c6299eedd0c77de82a47e9026f3762aee"} Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.141759 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"a2369a91497c650069e91175daa386bb571d406473471e9eb5624a8dbf181990"} Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.145172 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" event={"ID":"adab756d-1eb4-4837-a93a-346a38f83df9","Type":"ContainerStarted","Data":"c023c8982a0c0ad50ff5f66272882f7a5ffdc6ffbeba98d17209cd6023c76928"} Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.145198 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" event={"ID":"adab756d-1eb4-4837-a93a-346a38f83df9","Type":"ContainerStarted","Data":"e1954edfeedc886a54a667e9bc01c139c7487849899b090c4d5e26c2b0e662c2"} Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.145894 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.147193 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"4a6d9e2dbe862dbadeefd990c1e204c23f64c6676a3b7c7d957466f6f76af2a3"} Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.155966 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-2qdjc" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.171113 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" podStartSLOduration=130.171097047 podStartE2EDuration="2m10.171097047s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:14.169038741 +0000 UTC m=+148.488249410" watchObservedRunningTime="2025-12-09 15:34:14.171097047 +0000 UTC m=+148.490307716" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.384236 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q58q9"] Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.410925 4894 patch_prober.go:28] interesting pod/router-default-5444994796-sd4tb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 15:34:14 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 09 15:34:14 crc kubenswrapper[4894]: [+]process-running ok Dec 09 15:34:14 crc kubenswrapper[4894]: healthz check failed Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.411001 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sd4tb" podUID="a5600313-a000-42ce-94ca-898301d071fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.480922 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.482410 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 15:34:14 crc kubenswrapper[4894]: W1209 15:34:14.486556 4894 reflector.go:561] object-"openshift-kube-apiserver"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openshift-kube-apiserver": no relationship found between node 'crc' and this object Dec 09 15:34:14 crc kubenswrapper[4894]: E1209 15:34:14.486601 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-kube-apiserver\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openshift-kube-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:34:14 crc kubenswrapper[4894]: W1209 15:34:14.486666 4894 reflector.go:561] object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n": failed to list *v1.Secret: secrets "installer-sa-dockercfg-5pr6n" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-kube-apiserver": no relationship found between node 'crc' and this object Dec 09 15:34:14 crc kubenswrapper[4894]: E1209 15:34:14.486681 4894 reflector.go:158] "Unhandled Error" err="object-\"openshift-kube-apiserver\"/\"installer-sa-dockercfg-5pr6n\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"installer-sa-dockercfg-5pr6n\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-kube-apiserver\": no relationship found between node 'crc' and this object" logger="UnhandledError" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.498927 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4bsxc"] Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.500108 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bsxc" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.506043 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.522275 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.546137 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4bsxc"] Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.574513 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfb801bd-e785-4fdc-8a34-d2b8769d3a29-utilities\") pod \"redhat-operators-4bsxc\" (UID: \"bfb801bd-e785-4fdc-8a34-d2b8769d3a29\") " pod="openshift-marketplace/redhat-operators-4bsxc" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.574678 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b04fc35-284e-4150-a25b-715a2fa1399b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9b04fc35-284e-4150-a25b-715a2fa1399b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.574739 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-67ndf\" (UniqueName: \"kubernetes.io/projected/bfb801bd-e785-4fdc-8a34-d2b8769d3a29-kube-api-access-67ndf\") pod \"redhat-operators-4bsxc\" (UID: \"bfb801bd-e785-4fdc-8a34-d2b8769d3a29\") " pod="openshift-marketplace/redhat-operators-4bsxc" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.574781 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfb801bd-e785-4fdc-8a34-d2b8769d3a29-catalog-content\") pod \"redhat-operators-4bsxc\" (UID: \"bfb801bd-e785-4fdc-8a34-d2b8769d3a29\") " pod="openshift-marketplace/redhat-operators-4bsxc" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.574801 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9b04fc35-284e-4150-a25b-715a2fa1399b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9b04fc35-284e-4150-a25b-715a2fa1399b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.616433 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.676531 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfb801bd-e785-4fdc-8a34-d2b8769d3a29-utilities\") pod \"redhat-operators-4bsxc\" (UID: \"bfb801bd-e785-4fdc-8a34-d2b8769d3a29\") " pod="openshift-marketplace/redhat-operators-4bsxc" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.676610 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b04fc35-284e-4150-a25b-715a2fa1399b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9b04fc35-284e-4150-a25b-715a2fa1399b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.676678 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-67ndf\" (UniqueName: \"kubernetes.io/projected/bfb801bd-e785-4fdc-8a34-d2b8769d3a29-kube-api-access-67ndf\") pod \"redhat-operators-4bsxc\" (UID: \"bfb801bd-e785-4fdc-8a34-d2b8769d3a29\") " pod="openshift-marketplace/redhat-operators-4bsxc" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.676707 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfb801bd-e785-4fdc-8a34-d2b8769d3a29-catalog-content\") pod \"redhat-operators-4bsxc\" (UID: \"bfb801bd-e785-4fdc-8a34-d2b8769d3a29\") " pod="openshift-marketplace/redhat-operators-4bsxc" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.676731 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9b04fc35-284e-4150-a25b-715a2fa1399b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9b04fc35-284e-4150-a25b-715a2fa1399b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.676810 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9b04fc35-284e-4150-a25b-715a2fa1399b-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"9b04fc35-284e-4150-a25b-715a2fa1399b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.677119 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfb801bd-e785-4fdc-8a34-d2b8769d3a29-utilities\") pod \"redhat-operators-4bsxc\" (UID: \"bfb801bd-e785-4fdc-8a34-d2b8769d3a29\") " pod="openshift-marketplace/redhat-operators-4bsxc" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.677177 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfb801bd-e785-4fdc-8a34-d2b8769d3a29-catalog-content\") pod \"redhat-operators-4bsxc\" (UID: \"bfb801bd-e785-4fdc-8a34-d2b8769d3a29\") " pod="openshift-marketplace/redhat-operators-4bsxc" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.706747 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-67ndf\" (UniqueName: \"kubernetes.io/projected/bfb801bd-e785-4fdc-8a34-d2b8769d3a29-kube-api-access-67ndf\") pod \"redhat-operators-4bsxc\" (UID: \"bfb801bd-e785-4fdc-8a34-d2b8769d3a29\") " pod="openshift-marketplace/redhat-operators-4bsxc" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.723242 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.778110 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f67460f9-bc45-47e0-b186-f66dea9a578c-secret-volume\") pod \"f67460f9-bc45-47e0-b186-f66dea9a578c\" (UID: \"f67460f9-bc45-47e0-b186-f66dea9a578c\") " Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.778169 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f67460f9-bc45-47e0-b186-f66dea9a578c-config-volume\") pod \"f67460f9-bc45-47e0-b186-f66dea9a578c\" (UID: \"f67460f9-bc45-47e0-b186-f66dea9a578c\") " Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.778234 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jbvzc\" (UniqueName: \"kubernetes.io/projected/f67460f9-bc45-47e0-b186-f66dea9a578c-kube-api-access-jbvzc\") pod \"f67460f9-bc45-47e0-b186-f66dea9a578c\" (UID: \"f67460f9-bc45-47e0-b186-f66dea9a578c\") " Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.779904 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f67460f9-bc45-47e0-b186-f66dea9a578c-config-volume" (OuterVolumeSpecName: "config-volume") pod "f67460f9-bc45-47e0-b186-f66dea9a578c" (UID: "f67460f9-bc45-47e0-b186-f66dea9a578c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.782291 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f67460f9-bc45-47e0-b186-f66dea9a578c-kube-api-access-jbvzc" (OuterVolumeSpecName: "kube-api-access-jbvzc") pod "f67460f9-bc45-47e0-b186-f66dea9a578c" (UID: "f67460f9-bc45-47e0-b186-f66dea9a578c"). InnerVolumeSpecName "kube-api-access-jbvzc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.782728 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f67460f9-bc45-47e0-b186-f66dea9a578c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "f67460f9-bc45-47e0-b186-f66dea9a578c" (UID: "f67460f9-bc45-47e0-b186-f66dea9a578c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.845669 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bsxc" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.883860 4894 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/f67460f9-bc45-47e0-b186-f66dea9a578c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.883896 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jbvzc\" (UniqueName: \"kubernetes.io/projected/f67460f9-bc45-47e0-b186-f66dea9a578c-kube-api-access-jbvzc\") on node \"crc\" DevicePath \"\"" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.883908 4894 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/f67460f9-bc45-47e0-b186-f66dea9a578c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.889532 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-lz9jb"] Dec 09 15:34:14 crc kubenswrapper[4894]: E1209 15:34:14.889844 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f67460f9-bc45-47e0-b186-f66dea9a578c" containerName="collect-profiles" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.889869 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f67460f9-bc45-47e0-b186-f66dea9a578c" containerName="collect-profiles" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.889987 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f67460f9-bc45-47e0-b186-f66dea9a578c" containerName="collect-profiles" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.890788 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lz9jb" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.894681 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lz9jb"] Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.986329 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22822cfe-7b3b-4b3a-ba4f-c19630ccccae-catalog-content\") pod \"redhat-operators-lz9jb\" (UID: \"22822cfe-7b3b-4b3a-ba4f-c19630ccccae\") " pod="openshift-marketplace/redhat-operators-lz9jb" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.986404 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22822cfe-7b3b-4b3a-ba4f-c19630ccccae-utilities\") pod \"redhat-operators-lz9jb\" (UID: \"22822cfe-7b3b-4b3a-ba4f-c19630ccccae\") " pod="openshift-marketplace/redhat-operators-lz9jb" Dec 09 15:34:14 crc kubenswrapper[4894]: I1209 15:34:14.986465 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gckhc\" (UniqueName: \"kubernetes.io/projected/22822cfe-7b3b-4b3a-ba4f-c19630ccccae-kube-api-access-gckhc\") pod \"redhat-operators-lz9jb\" (UID: \"22822cfe-7b3b-4b3a-ba4f-c19630ccccae\") " pod="openshift-marketplace/redhat-operators-lz9jb" Dec 09 15:34:15 crc kubenswrapper[4894]: I1209 15:34:15.091299 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22822cfe-7b3b-4b3a-ba4f-c19630ccccae-utilities\") pod \"redhat-operators-lz9jb\" (UID: \"22822cfe-7b3b-4b3a-ba4f-c19630ccccae\") " pod="openshift-marketplace/redhat-operators-lz9jb" Dec 09 15:34:15 crc kubenswrapper[4894]: I1209 15:34:15.091421 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gckhc\" (UniqueName: \"kubernetes.io/projected/22822cfe-7b3b-4b3a-ba4f-c19630ccccae-kube-api-access-gckhc\") pod \"redhat-operators-lz9jb\" (UID: \"22822cfe-7b3b-4b3a-ba4f-c19630ccccae\") " pod="openshift-marketplace/redhat-operators-lz9jb" Dec 09 15:34:15 crc kubenswrapper[4894]: I1209 15:34:15.091456 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22822cfe-7b3b-4b3a-ba4f-c19630ccccae-catalog-content\") pod \"redhat-operators-lz9jb\" (UID: \"22822cfe-7b3b-4b3a-ba4f-c19630ccccae\") " pod="openshift-marketplace/redhat-operators-lz9jb" Dec 09 15:34:15 crc kubenswrapper[4894]: I1209 15:34:15.092242 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22822cfe-7b3b-4b3a-ba4f-c19630ccccae-utilities\") pod \"redhat-operators-lz9jb\" (UID: \"22822cfe-7b3b-4b3a-ba4f-c19630ccccae\") " pod="openshift-marketplace/redhat-operators-lz9jb" Dec 09 15:34:15 crc kubenswrapper[4894]: I1209 15:34:15.092278 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22822cfe-7b3b-4b3a-ba4f-c19630ccccae-catalog-content\") pod \"redhat-operators-lz9jb\" (UID: \"22822cfe-7b3b-4b3a-ba4f-c19630ccccae\") " pod="openshift-marketplace/redhat-operators-lz9jb" Dec 09 15:34:15 crc kubenswrapper[4894]: I1209 15:34:15.122179 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gckhc\" (UniqueName: \"kubernetes.io/projected/22822cfe-7b3b-4b3a-ba4f-c19630ccccae-kube-api-access-gckhc\") pod \"redhat-operators-lz9jb\" (UID: \"22822cfe-7b3b-4b3a-ba4f-c19630ccccae\") " pod="openshift-marketplace/redhat-operators-lz9jb" Dec 09 15:34:15 crc kubenswrapper[4894]: I1209 15:34:15.126312 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4bsxc"] Dec 09 15:34:15 crc kubenswrapper[4894]: W1209 15:34:15.141911 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbfb801bd_e785_4fdc_8a34_d2b8769d3a29.slice/crio-425abcd653efcfd1107b3fd4179cf301058483398ea6aebdccf719c14c17a786 WatchSource:0}: Error finding container 425abcd653efcfd1107b3fd4179cf301058483398ea6aebdccf719c14c17a786: Status 404 returned error can't find the container with id 425abcd653efcfd1107b3fd4179cf301058483398ea6aebdccf719c14c17a786 Dec 09 15:34:15 crc kubenswrapper[4894]: I1209 15:34:15.153494 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bsxc" event={"ID":"bfb801bd-e785-4fdc-8a34-d2b8769d3a29","Type":"ContainerStarted","Data":"425abcd653efcfd1107b3fd4179cf301058483398ea6aebdccf719c14c17a786"} Dec 09 15:34:15 crc kubenswrapper[4894]: I1209 15:34:15.155735 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q58q9" event={"ID":"ee374eee-998a-48f2-acdf-39daa5ce2fdd","Type":"ContainerStarted","Data":"b2bf84540592cbaa8843f5b8d3839eebf0d3a1fb44ef2be442cffd16639cd089"} Dec 09 15:34:15 crc kubenswrapper[4894]: I1209 15:34:15.157015 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5832b82d-0bc8-434a-9907-7554ca34a3a6","Type":"ContainerStarted","Data":"131b777159873a8f7214e9c4a5c956d36515aab3a2e398f7295fed8153685169"} Dec 09 15:34:15 crc kubenswrapper[4894]: I1209 15:34:15.158399 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r" event={"ID":"f67460f9-bc45-47e0-b186-f66dea9a578c","Type":"ContainerDied","Data":"8d0fe7df17a05f02f99d71e792190b18f61f1493f9d1a5f9b393d19d743f0d6b"} Dec 09 15:34:15 crc kubenswrapper[4894]: I1209 15:34:15.158429 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8d0fe7df17a05f02f99d71e792190b18f61f1493f9d1a5f9b393d19d743f0d6b" Dec 09 15:34:15 crc kubenswrapper[4894]: I1209 15:34:15.158595 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r" Dec 09 15:34:15 crc kubenswrapper[4894]: I1209 15:34:15.242937 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lz9jb" Dec 09 15:34:15 crc kubenswrapper[4894]: I1209 15:34:15.392887 4894 patch_prober.go:28] interesting pod/router-default-5444994796-sd4tb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 15:34:15 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 09 15:34:15 crc kubenswrapper[4894]: [+]process-running ok Dec 09 15:34:15 crc kubenswrapper[4894]: healthz check failed Dec 09 15:34:15 crc kubenswrapper[4894]: I1209 15:34:15.392950 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sd4tb" podUID="a5600313-a000-42ce-94ca-898301d071fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:34:15 crc kubenswrapper[4894]: I1209 15:34:15.461189 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 09 15:34:15 crc kubenswrapper[4894]: I1209 15:34:15.477227 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b04fc35-284e-4150-a25b-715a2fa1399b-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"9b04fc35-284e-4150-a25b-715a2fa1399b\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 15:34:15 crc kubenswrapper[4894]: I1209 15:34:15.543185 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-lz9jb"] Dec 09 15:34:15 crc kubenswrapper[4894]: W1209 15:34:15.552428 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod22822cfe_7b3b_4b3a_ba4f_c19630ccccae.slice/crio-02255fc0367f25e7867f62c6f0b2589b76b9b10ba9d943abfad2db2644cee693 WatchSource:0}: Error finding container 02255fc0367f25e7867f62c6f0b2589b76b9b10ba9d943abfad2db2644cee693: Status 404 returned error can't find the container with id 02255fc0367f25e7867f62c6f0b2589b76b9b10ba9d943abfad2db2644cee693 Dec 09 15:34:15 crc kubenswrapper[4894]: I1209 15:34:15.754372 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 09 15:34:15 crc kubenswrapper[4894]: I1209 15:34:15.763693 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 15:34:16 crc kubenswrapper[4894]: I1209 15:34:16.180106 4894 generic.go:334] "Generic (PLEG): container finished" podID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" containerID="145672364d83611684af8e281cb8d9a14305f05659503f09298059c0eb54a7d1" exitCode=0 Dec 09 15:34:16 crc kubenswrapper[4894]: I1209 15:34:16.180256 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bsxc" event={"ID":"bfb801bd-e785-4fdc-8a34-d2b8769d3a29","Type":"ContainerDied","Data":"145672364d83611684af8e281cb8d9a14305f05659503f09298059c0eb54a7d1"} Dec 09 15:34:16 crc kubenswrapper[4894]: I1209 15:34:16.183457 4894 generic.go:334] "Generic (PLEG): container finished" podID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" containerID="9ade3816e5bf9bba5a076fc5ba04468b688cdbcb09e029720868c42858aecfff" exitCode=0 Dec 09 15:34:16 crc kubenswrapper[4894]: I1209 15:34:16.183517 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q58q9" event={"ID":"ee374eee-998a-48f2-acdf-39daa5ce2fdd","Type":"ContainerDied","Data":"9ade3816e5bf9bba5a076fc5ba04468b688cdbcb09e029720868c42858aecfff"} Dec 09 15:34:16 crc kubenswrapper[4894]: I1209 15:34:16.190185 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"cfffcd2aa8c7313734409c4b0a7ce16a676912d91ddfa715f129afca600ef0c6"} Dec 09 15:34:16 crc kubenswrapper[4894]: I1209 15:34:16.190264 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:34:16 crc kubenswrapper[4894]: I1209 15:34:16.193023 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5832b82d-0bc8-434a-9907-7554ca34a3a6","Type":"ContainerStarted","Data":"94be6d016c7780c36a364c6a21a91e6484d858573461683c0fbaeb4354ee4f05"} Dec 09 15:34:16 crc kubenswrapper[4894]: I1209 15:34:16.195211 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"1ec3bc8b6a6fe93480a27cdc668739743930cde584f10da92e029f6079a29946"} Dec 09 15:34:16 crc kubenswrapper[4894]: I1209 15:34:16.198140 4894 generic.go:334] "Generic (PLEG): container finished" podID="22822cfe-7b3b-4b3a-ba4f-c19630ccccae" containerID="52c5871e23c7392d33bdc1375b7d6206b816af32b465ffcd85d034674e5c4893" exitCode=0 Dec 09 15:34:16 crc kubenswrapper[4894]: I1209 15:34:16.198194 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lz9jb" event={"ID":"22822cfe-7b3b-4b3a-ba4f-c19630ccccae","Type":"ContainerDied","Data":"52c5871e23c7392d33bdc1375b7d6206b816af32b465ffcd85d034674e5c4893"} Dec 09 15:34:16 crc kubenswrapper[4894]: I1209 15:34:16.198211 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lz9jb" event={"ID":"22822cfe-7b3b-4b3a-ba4f-c19630ccccae","Type":"ContainerStarted","Data":"02255fc0367f25e7867f62c6f0b2589b76b9b10ba9d943abfad2db2644cee693"} Dec 09 15:34:16 crc kubenswrapper[4894]: I1209 15:34:16.207706 4894 generic.go:334] "Generic (PLEG): container finished" podID="2421584b-8a4b-4cb1-97c0-cb41d8ec16b2" containerID="6356f5ee31ced0642b88fc9d15acdb1b7ce35ffee04c3e156a38a54f45ac947c" exitCode=0 Dec 09 15:34:16 crc kubenswrapper[4894]: I1209 15:34:16.207790 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxlqt" event={"ID":"2421584b-8a4b-4cb1-97c0-cb41d8ec16b2","Type":"ContainerDied","Data":"6356f5ee31ced0642b88fc9d15acdb1b7ce35ffee04c3e156a38a54f45ac947c"} Dec 09 15:34:16 crc kubenswrapper[4894]: I1209 15:34:16.221147 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"40ffdddf4ec8c62865699b8385a96f77598aabee20ccc2189ccebf42410b59f0"} Dec 09 15:34:16 crc kubenswrapper[4894]: I1209 15:34:16.291184 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Dec 09 15:34:16 crc kubenswrapper[4894]: W1209 15:34:16.332353 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod9b04fc35_284e_4150_a25b_715a2fa1399b.slice/crio-aaf3cef521c43d3a744b528347e6efc9b3bf5724c4d7b9d9d3d39498c2404b11 WatchSource:0}: Error finding container aaf3cef521c43d3a744b528347e6efc9b3bf5724c4d7b9d9d3d39498c2404b11: Status 404 returned error can't find the container with id aaf3cef521c43d3a744b528347e6efc9b3bf5724c4d7b9d9d3d39498c2404b11 Dec 09 15:34:16 crc kubenswrapper[4894]: I1209 15:34:16.401589 4894 patch_prober.go:28] interesting pod/router-default-5444994796-sd4tb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 15:34:16 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 09 15:34:16 crc kubenswrapper[4894]: [+]process-running ok Dec 09 15:34:16 crc kubenswrapper[4894]: healthz check failed Dec 09 15:34:16 crc kubenswrapper[4894]: I1209 15:34:16.401770 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sd4tb" podUID="a5600313-a000-42ce-94ca-898301d071fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:34:16 crc kubenswrapper[4894]: I1209 15:34:16.464099 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=3.464073814 podStartE2EDuration="3.464073814s" podCreationTimestamp="2025-12-09 15:34:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:16.460309883 +0000 UTC m=+150.779520552" watchObservedRunningTime="2025-12-09 15:34:16.464073814 +0000 UTC m=+150.783284483" Dec 09 15:34:17 crc kubenswrapper[4894]: I1209 15:34:17.234834 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9b04fc35-284e-4150-a25b-715a2fa1399b","Type":"ContainerStarted","Data":"1d93b362d4607dd4767c417079844229743bf4726cd11deb57062d6cdda34584"} Dec 09 15:34:17 crc kubenswrapper[4894]: I1209 15:34:17.235191 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9b04fc35-284e-4150-a25b-715a2fa1399b","Type":"ContainerStarted","Data":"aaf3cef521c43d3a744b528347e6efc9b3bf5724c4d7b9d9d3d39498c2404b11"} Dec 09 15:34:17 crc kubenswrapper[4894]: I1209 15:34:17.239411 4894 generic.go:334] "Generic (PLEG): container finished" podID="5832b82d-0bc8-434a-9907-7554ca34a3a6" containerID="94be6d016c7780c36a364c6a21a91e6484d858573461683c0fbaeb4354ee4f05" exitCode=0 Dec 09 15:34:17 crc kubenswrapper[4894]: I1209 15:34:17.239555 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5832b82d-0bc8-434a-9907-7554ca34a3a6","Type":"ContainerDied","Data":"94be6d016c7780c36a364c6a21a91e6484d858573461683c0fbaeb4354ee4f05"} Dec 09 15:34:17 crc kubenswrapper[4894]: I1209 15:34:17.252256 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=3.252237529 podStartE2EDuration="3.252237529s" podCreationTimestamp="2025-12-09 15:34:14 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:34:17.251420247 +0000 UTC m=+151.570630916" watchObservedRunningTime="2025-12-09 15:34:17.252237529 +0000 UTC m=+151.571448198" Dec 09 15:34:17 crc kubenswrapper[4894]: I1209 15:34:17.388711 4894 patch_prober.go:28] interesting pod/router-default-5444994796-sd4tb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 15:34:17 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 09 15:34:17 crc kubenswrapper[4894]: [+]process-running ok Dec 09 15:34:17 crc kubenswrapper[4894]: healthz check failed Dec 09 15:34:17 crc kubenswrapper[4894]: I1209 15:34:17.388781 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sd4tb" podUID="a5600313-a000-42ce-94ca-898301d071fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:34:18 crc kubenswrapper[4894]: I1209 15:34:18.275286 4894 generic.go:334] "Generic (PLEG): container finished" podID="9b04fc35-284e-4150-a25b-715a2fa1399b" containerID="1d93b362d4607dd4767c417079844229743bf4726cd11deb57062d6cdda34584" exitCode=0 Dec 09 15:34:18 crc kubenswrapper[4894]: I1209 15:34:18.275371 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9b04fc35-284e-4150-a25b-715a2fa1399b","Type":"ContainerDied","Data":"1d93b362d4607dd4767c417079844229743bf4726cd11deb57062d6cdda34584"} Dec 09 15:34:18 crc kubenswrapper[4894]: I1209 15:34:18.393106 4894 patch_prober.go:28] interesting pod/router-default-5444994796-sd4tb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 15:34:18 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 09 15:34:18 crc kubenswrapper[4894]: [+]process-running ok Dec 09 15:34:18 crc kubenswrapper[4894]: healthz check failed Dec 09 15:34:18 crc kubenswrapper[4894]: I1209 15:34:18.393160 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sd4tb" podUID="a5600313-a000-42ce-94ca-898301d071fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:34:18 crc kubenswrapper[4894]: I1209 15:34:18.505185 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-d5smt" Dec 09 15:34:18 crc kubenswrapper[4894]: I1209 15:34:18.833970 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 15:34:18 crc kubenswrapper[4894]: I1209 15:34:18.919864 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5832b82d-0bc8-434a-9907-7554ca34a3a6-kube-api-access\") pod \"5832b82d-0bc8-434a-9907-7554ca34a3a6\" (UID: \"5832b82d-0bc8-434a-9907-7554ca34a3a6\") " Dec 09 15:34:18 crc kubenswrapper[4894]: I1209 15:34:18.919972 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5832b82d-0bc8-434a-9907-7554ca34a3a6-kubelet-dir\") pod \"5832b82d-0bc8-434a-9907-7554ca34a3a6\" (UID: \"5832b82d-0bc8-434a-9907-7554ca34a3a6\") " Dec 09 15:34:18 crc kubenswrapper[4894]: I1209 15:34:18.920069 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5832b82d-0bc8-434a-9907-7554ca34a3a6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5832b82d-0bc8-434a-9907-7554ca34a3a6" (UID: "5832b82d-0bc8-434a-9907-7554ca34a3a6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:34:18 crc kubenswrapper[4894]: I1209 15:34:18.920535 4894 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5832b82d-0bc8-434a-9907-7554ca34a3a6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 09 15:34:18 crc kubenswrapper[4894]: I1209 15:34:18.930584 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5832b82d-0bc8-434a-9907-7554ca34a3a6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5832b82d-0bc8-434a-9907-7554ca34a3a6" (UID: "5832b82d-0bc8-434a-9907-7554ca34a3a6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:34:19 crc kubenswrapper[4894]: I1209 15:34:19.005233 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:19 crc kubenswrapper[4894]: I1209 15:34:19.011702 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-bcfbr" Dec 09 15:34:19 crc kubenswrapper[4894]: I1209 15:34:19.028767 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5832b82d-0bc8-434a-9907-7554ca34a3a6-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 15:34:19 crc kubenswrapper[4894]: I1209 15:34:19.309667 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"5832b82d-0bc8-434a-9907-7554ca34a3a6","Type":"ContainerDied","Data":"131b777159873a8f7214e9c4a5c956d36515aab3a2e398f7295fed8153685169"} Dec 09 15:34:19 crc kubenswrapper[4894]: I1209 15:34:19.309740 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="131b777159873a8f7214e9c4a5c956d36515aab3a2e398f7295fed8153685169" Dec 09 15:34:19 crc kubenswrapper[4894]: I1209 15:34:19.309883 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Dec 09 15:34:19 crc kubenswrapper[4894]: I1209 15:34:19.388748 4894 patch_prober.go:28] interesting pod/router-default-5444994796-sd4tb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 15:34:19 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 09 15:34:19 crc kubenswrapper[4894]: [+]process-running ok Dec 09 15:34:19 crc kubenswrapper[4894]: healthz check failed Dec 09 15:34:19 crc kubenswrapper[4894]: I1209 15:34:19.388807 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sd4tb" podUID="a5600313-a000-42ce-94ca-898301d071fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:34:19 crc kubenswrapper[4894]: I1209 15:34:19.788958 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 15:34:19 crc kubenswrapper[4894]: I1209 15:34:19.861267 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b04fc35-284e-4150-a25b-715a2fa1399b-kube-api-access\") pod \"9b04fc35-284e-4150-a25b-715a2fa1399b\" (UID: \"9b04fc35-284e-4150-a25b-715a2fa1399b\") " Dec 09 15:34:19 crc kubenswrapper[4894]: I1209 15:34:19.861691 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9b04fc35-284e-4150-a25b-715a2fa1399b-kubelet-dir\") pod \"9b04fc35-284e-4150-a25b-715a2fa1399b\" (UID: \"9b04fc35-284e-4150-a25b-715a2fa1399b\") " Dec 09 15:34:19 crc kubenswrapper[4894]: I1209 15:34:19.861805 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9b04fc35-284e-4150-a25b-715a2fa1399b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "9b04fc35-284e-4150-a25b-715a2fa1399b" (UID: "9b04fc35-284e-4150-a25b-715a2fa1399b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:34:19 crc kubenswrapper[4894]: I1209 15:34:19.862009 4894 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/9b04fc35-284e-4150-a25b-715a2fa1399b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 09 15:34:19 crc kubenswrapper[4894]: I1209 15:34:19.872202 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b04fc35-284e-4150-a25b-715a2fa1399b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "9b04fc35-284e-4150-a25b-715a2fa1399b" (UID: "9b04fc35-284e-4150-a25b-715a2fa1399b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:34:19 crc kubenswrapper[4894]: I1209 15:34:19.962947 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/9b04fc35-284e-4150-a25b-715a2fa1399b-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 15:34:20 crc kubenswrapper[4894]: I1209 15:34:20.324053 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"9b04fc35-284e-4150-a25b-715a2fa1399b","Type":"ContainerDied","Data":"aaf3cef521c43d3a744b528347e6efc9b3bf5724c4d7b9d9d3d39498c2404b11"} Dec 09 15:34:20 crc kubenswrapper[4894]: I1209 15:34:20.324088 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aaf3cef521c43d3a744b528347e6efc9b3bf5724c4d7b9d9d3d39498c2404b11" Dec 09 15:34:20 crc kubenswrapper[4894]: I1209 15:34:20.324139 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Dec 09 15:34:20 crc kubenswrapper[4894]: I1209 15:34:20.392447 4894 patch_prober.go:28] interesting pod/router-default-5444994796-sd4tb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 15:34:20 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 09 15:34:20 crc kubenswrapper[4894]: [+]process-running ok Dec 09 15:34:20 crc kubenswrapper[4894]: healthz check failed Dec 09 15:34:20 crc kubenswrapper[4894]: I1209 15:34:20.392495 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sd4tb" podUID="a5600313-a000-42ce-94ca-898301d071fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:34:21 crc kubenswrapper[4894]: I1209 15:34:21.390022 4894 patch_prober.go:28] interesting pod/router-default-5444994796-sd4tb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 15:34:21 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 09 15:34:21 crc kubenswrapper[4894]: [+]process-running ok Dec 09 15:34:21 crc kubenswrapper[4894]: healthz check failed Dec 09 15:34:21 crc kubenswrapper[4894]: I1209 15:34:21.390092 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sd4tb" podUID="a5600313-a000-42ce-94ca-898301d071fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:34:21 crc kubenswrapper[4894]: I1209 15:34:21.679969 4894 patch_prober.go:28] interesting pod/console-f9d7485db-crhvg container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Dec 09 15:34:21 crc kubenswrapper[4894]: I1209 15:34:21.680044 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-crhvg" podUID="9d548d70-c4af-4333-8cea-f6bef5722ce5" containerName="console" probeResult="failure" output="Get \"https://10.217.0.11:8443/health\": dial tcp 10.217.0.11:8443: connect: connection refused" Dec 09 15:34:22 crc kubenswrapper[4894]: I1209 15:34:22.303582 4894 patch_prober.go:28] interesting pod/downloads-7954f5f757-vtbc8 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 09 15:34:22 crc kubenswrapper[4894]: I1209 15:34:22.303974 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-vtbc8" podUID="caed7380-e514-4f40-ae98-0be4a2fe61c3" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 09 15:34:22 crc kubenswrapper[4894]: I1209 15:34:22.303828 4894 patch_prober.go:28] interesting pod/downloads-7954f5f757-vtbc8 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" start-of-body= Dec 09 15:34:22 crc kubenswrapper[4894]: I1209 15:34:22.304327 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-vtbc8" podUID="caed7380-e514-4f40-ae98-0be4a2fe61c3" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.14:8080/\": dial tcp 10.217.0.14:8080: connect: connection refused" Dec 09 15:34:22 crc kubenswrapper[4894]: I1209 15:34:22.391850 4894 patch_prober.go:28] interesting pod/router-default-5444994796-sd4tb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 15:34:22 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 09 15:34:22 crc kubenswrapper[4894]: [+]process-running ok Dec 09 15:34:22 crc kubenswrapper[4894]: healthz check failed Dec 09 15:34:22 crc kubenswrapper[4894]: I1209 15:34:22.391937 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sd4tb" podUID="a5600313-a000-42ce-94ca-898301d071fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:34:23 crc kubenswrapper[4894]: I1209 15:34:23.389935 4894 patch_prober.go:28] interesting pod/router-default-5444994796-sd4tb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 15:34:23 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 09 15:34:23 crc kubenswrapper[4894]: [+]process-running ok Dec 09 15:34:23 crc kubenswrapper[4894]: healthz check failed Dec 09 15:34:23 crc kubenswrapper[4894]: I1209 15:34:23.390040 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sd4tb" podUID="a5600313-a000-42ce-94ca-898301d071fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:34:24 crc kubenswrapper[4894]: I1209 15:34:24.388014 4894 patch_prober.go:28] interesting pod/router-default-5444994796-sd4tb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 15:34:24 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 09 15:34:24 crc kubenswrapper[4894]: [+]process-running ok Dec 09 15:34:24 crc kubenswrapper[4894]: healthz check failed Dec 09 15:34:24 crc kubenswrapper[4894]: I1209 15:34:24.388327 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sd4tb" podUID="a5600313-a000-42ce-94ca-898301d071fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:34:25 crc kubenswrapper[4894]: I1209 15:34:25.390853 4894 patch_prober.go:28] interesting pod/router-default-5444994796-sd4tb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 15:34:25 crc kubenswrapper[4894]: [-]has-synced failed: reason withheld Dec 09 15:34:25 crc kubenswrapper[4894]: [+]process-running ok Dec 09 15:34:25 crc kubenswrapper[4894]: healthz check failed Dec 09 15:34:25 crc kubenswrapper[4894]: I1209 15:34:25.390903 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sd4tb" podUID="a5600313-a000-42ce-94ca-898301d071fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:34:26 crc kubenswrapper[4894]: I1209 15:34:26.393279 4894 patch_prober.go:28] interesting pod/router-default-5444994796-sd4tb container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Dec 09 15:34:26 crc kubenswrapper[4894]: [+]has-synced ok Dec 09 15:34:26 crc kubenswrapper[4894]: [+]process-running ok Dec 09 15:34:26 crc kubenswrapper[4894]: healthz check failed Dec 09 15:34:26 crc kubenswrapper[4894]: I1209 15:34:26.393359 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-sd4tb" podUID="a5600313-a000-42ce-94ca-898301d071fd" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:34:26 crc kubenswrapper[4894]: I1209 15:34:26.893430 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs\") pod \"network-metrics-daemon-qxfr8\" (UID: \"948b28c5-e35c-4e6c-b31c-c0d5b205389b\") " pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:34:26 crc kubenswrapper[4894]: I1209 15:34:26.918795 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/948b28c5-e35c-4e6c-b31c-c0d5b205389b-metrics-certs\") pod \"network-metrics-daemon-qxfr8\" (UID: \"948b28c5-e35c-4e6c-b31c-c0d5b205389b\") " pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:34:27 crc kubenswrapper[4894]: I1209 15:34:27.133860 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-qxfr8" Dec 09 15:34:27 crc kubenswrapper[4894]: I1209 15:34:27.390055 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-sd4tb" Dec 09 15:34:27 crc kubenswrapper[4894]: I1209 15:34:27.392456 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-sd4tb" Dec 09 15:34:32 crc kubenswrapper[4894]: I1209 15:34:32.309855 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-vtbc8" Dec 09 15:34:32 crc kubenswrapper[4894]: I1209 15:34:32.330227 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:34:32 crc kubenswrapper[4894]: I1209 15:34:32.337505 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:34:32 crc kubenswrapper[4894]: I1209 15:34:32.534142 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:34:42 crc kubenswrapper[4894]: I1209 15:34:42.258022 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:34:42 crc kubenswrapper[4894]: I1209 15:34:42.258387 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:34:43 crc kubenswrapper[4894]: I1209 15:34:43.683012 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-m8gtv" Dec 09 15:34:48 crc kubenswrapper[4894]: I1209 15:34:48.850306 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 09 15:34:48 crc kubenswrapper[4894]: E1209 15:34:48.850808 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b04fc35-284e-4150-a25b-715a2fa1399b" containerName="pruner" Dec 09 15:34:48 crc kubenswrapper[4894]: I1209 15:34:48.850820 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b04fc35-284e-4150-a25b-715a2fa1399b" containerName="pruner" Dec 09 15:34:48 crc kubenswrapper[4894]: E1209 15:34:48.850827 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5832b82d-0bc8-434a-9907-7554ca34a3a6" containerName="pruner" Dec 09 15:34:48 crc kubenswrapper[4894]: I1209 15:34:48.850833 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5832b82d-0bc8-434a-9907-7554ca34a3a6" containerName="pruner" Dec 09 15:34:48 crc kubenswrapper[4894]: I1209 15:34:48.850960 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b04fc35-284e-4150-a25b-715a2fa1399b" containerName="pruner" Dec 09 15:34:48 crc kubenswrapper[4894]: I1209 15:34:48.850971 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="5832b82d-0bc8-434a-9907-7554ca34a3a6" containerName="pruner" Dec 09 15:34:48 crc kubenswrapper[4894]: I1209 15:34:48.851370 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 15:34:48 crc kubenswrapper[4894]: I1209 15:34:48.854884 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Dec 09 15:34:48 crc kubenswrapper[4894]: I1209 15:34:48.855835 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Dec 09 15:34:48 crc kubenswrapper[4894]: I1209 15:34:48.860942 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 09 15:34:48 crc kubenswrapper[4894]: I1209 15:34:48.933427 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a72fe2b-51af-49e9-a59f-de142120033b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0a72fe2b-51af-49e9-a59f-de142120033b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 15:34:48 crc kubenswrapper[4894]: I1209 15:34:48.933483 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0a72fe2b-51af-49e9-a59f-de142120033b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0a72fe2b-51af-49e9-a59f-de142120033b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 15:34:49 crc kubenswrapper[4894]: I1209 15:34:49.034941 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0a72fe2b-51af-49e9-a59f-de142120033b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0a72fe2b-51af-49e9-a59f-de142120033b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 15:34:49 crc kubenswrapper[4894]: I1209 15:34:49.035081 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0a72fe2b-51af-49e9-a59f-de142120033b-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"0a72fe2b-51af-49e9-a59f-de142120033b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 15:34:49 crc kubenswrapper[4894]: I1209 15:34:49.035094 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a72fe2b-51af-49e9-a59f-de142120033b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0a72fe2b-51af-49e9-a59f-de142120033b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 15:34:49 crc kubenswrapper[4894]: I1209 15:34:49.074483 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a72fe2b-51af-49e9-a59f-de142120033b-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"0a72fe2b-51af-49e9-a59f-de142120033b\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 15:34:49 crc kubenswrapper[4894]: I1209 15:34:49.185083 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 15:34:53 crc kubenswrapper[4894]: I1209 15:34:53.406453 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Dec 09 15:34:53 crc kubenswrapper[4894]: I1209 15:34:53.859125 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 09 15:34:53 crc kubenswrapper[4894]: I1209 15:34:53.859810 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 09 15:34:53 crc kubenswrapper[4894]: I1209 15:34:53.869217 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 09 15:34:53 crc kubenswrapper[4894]: I1209 15:34:53.903319 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5498bfbc-4945-4ffc-a7e3-db254f6cbabb-kube-api-access\") pod \"installer-9-crc\" (UID: \"5498bfbc-4945-4ffc-a7e3-db254f6cbabb\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 15:34:53 crc kubenswrapper[4894]: I1209 15:34:53.903371 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5498bfbc-4945-4ffc-a7e3-db254f6cbabb-kubelet-dir\") pod \"installer-9-crc\" (UID: \"5498bfbc-4945-4ffc-a7e3-db254f6cbabb\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 15:34:53 crc kubenswrapper[4894]: I1209 15:34:53.903390 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/5498bfbc-4945-4ffc-a7e3-db254f6cbabb-var-lock\") pod \"installer-9-crc\" (UID: \"5498bfbc-4945-4ffc-a7e3-db254f6cbabb\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 15:34:54 crc kubenswrapper[4894]: I1209 15:34:54.004968 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5498bfbc-4945-4ffc-a7e3-db254f6cbabb-kube-api-access\") pod \"installer-9-crc\" (UID: \"5498bfbc-4945-4ffc-a7e3-db254f6cbabb\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 15:34:54 crc kubenswrapper[4894]: I1209 15:34:54.005344 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5498bfbc-4945-4ffc-a7e3-db254f6cbabb-kubelet-dir\") pod \"installer-9-crc\" (UID: \"5498bfbc-4945-4ffc-a7e3-db254f6cbabb\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 15:34:54 crc kubenswrapper[4894]: I1209 15:34:54.005369 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/5498bfbc-4945-4ffc-a7e3-db254f6cbabb-var-lock\") pod \"installer-9-crc\" (UID: \"5498bfbc-4945-4ffc-a7e3-db254f6cbabb\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 15:34:54 crc kubenswrapper[4894]: I1209 15:34:54.005489 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/5498bfbc-4945-4ffc-a7e3-db254f6cbabb-var-lock\") pod \"installer-9-crc\" (UID: \"5498bfbc-4945-4ffc-a7e3-db254f6cbabb\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 15:34:54 crc kubenswrapper[4894]: I1209 15:34:54.005487 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5498bfbc-4945-4ffc-a7e3-db254f6cbabb-kubelet-dir\") pod \"installer-9-crc\" (UID: \"5498bfbc-4945-4ffc-a7e3-db254f6cbabb\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 15:34:54 crc kubenswrapper[4894]: I1209 15:34:54.058931 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5498bfbc-4945-4ffc-a7e3-db254f6cbabb-kube-api-access\") pod \"installer-9-crc\" (UID: \"5498bfbc-4945-4ffc-a7e3-db254f6cbabb\") " pod="openshift-kube-apiserver/installer-9-crc" Dec 09 15:34:54 crc kubenswrapper[4894]: I1209 15:34:54.181298 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 09 15:34:57 crc kubenswrapper[4894]: E1209 15:34:57.794748 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 09 15:34:57 crc kubenswrapper[4894]: E1209 15:34:57.795125 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gckhc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-lz9jb_openshift-marketplace(22822cfe-7b3b-4b3a-ba4f-c19630ccccae): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 15:34:57 crc kubenswrapper[4894]: E1209 15:34:57.796345 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-lz9jb" podUID="22822cfe-7b3b-4b3a-ba4f-c19630ccccae" Dec 09 15:34:59 crc kubenswrapper[4894]: E1209 15:34:59.370423 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-lz9jb" podUID="22822cfe-7b3b-4b3a-ba4f-c19630ccccae" Dec 09 15:34:59 crc kubenswrapper[4894]: E1209 15:34:59.459518 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 09 15:34:59 crc kubenswrapper[4894]: E1209 15:34:59.459932 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gqpnd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-fxq7l_openshift-marketplace(e042f2bf-1069-42a4-aef2-72ec5da57c99): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 15:34:59 crc kubenswrapper[4894]: E1209 15:34:59.461198 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-fxq7l" podUID="e042f2bf-1069-42a4-aef2-72ec5da57c99" Dec 09 15:35:00 crc kubenswrapper[4894]: E1209 15:35:00.835895 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-fxq7l" podUID="e042f2bf-1069-42a4-aef2-72ec5da57c99" Dec 09 15:35:00 crc kubenswrapper[4894]: E1209 15:35:00.898695 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 09 15:35:00 crc kubenswrapper[4894]: E1209 15:35:00.898927 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-r4d2d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-ff286_openshift-marketplace(c278157d-d806-4fa1-a597-076d95a286e2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 15:35:00 crc kubenswrapper[4894]: E1209 15:35:00.900113 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-ff286" podUID="c278157d-d806-4fa1-a597-076d95a286e2" Dec 09 15:35:00 crc kubenswrapper[4894]: E1209 15:35:00.918804 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Dec 09 15:35:00 crc kubenswrapper[4894]: E1209 15:35:00.918956 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4k9sg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-hq8h6_openshift-marketplace(4397f08d-278a-4e2d-9664-e44161c96638): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 15:35:00 crc kubenswrapper[4894]: E1209 15:35:00.921075 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-hq8h6" podUID="4397f08d-278a-4e2d-9664-e44161c96638" Dec 09 15:35:01 crc kubenswrapper[4894]: E1209 15:35:01.893802 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-ff286" podUID="c278157d-d806-4fa1-a597-076d95a286e2" Dec 09 15:35:01 crc kubenswrapper[4894]: E1209 15:35:01.893816 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-hq8h6" podUID="4397f08d-278a-4e2d-9664-e44161c96638" Dec 09 15:35:01 crc kubenswrapper[4894]: E1209 15:35:01.958252 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 09 15:35:01 crc kubenswrapper[4894]: E1209 15:35:01.958787 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-whpnq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-nxlqt_openshift-marketplace(2421584b-8a4b-4cb1-97c0-cb41d8ec16b2): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 15:35:01 crc kubenswrapper[4894]: E1209 15:35:01.960321 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-nxlqt" podUID="2421584b-8a4b-4cb1-97c0-cb41d8ec16b2" Dec 09 15:35:01 crc kubenswrapper[4894]: E1209 15:35:01.983609 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Dec 09 15:35:01 crc kubenswrapper[4894]: E1209 15:35:01.983782 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-trr6z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-s4rbb_openshift-marketplace(46d076e7-ca39-4337-bacc-ea3984f3bb5d): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 15:35:01 crc kubenswrapper[4894]: E1209 15:35:01.985188 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-s4rbb" podUID="46d076e7-ca39-4337-bacc-ea3984f3bb5d" Dec 09 15:35:01 crc kubenswrapper[4894]: E1209 15:35:01.997084 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Dec 09 15:35:01 crc kubenswrapper[4894]: E1209 15:35:01.997211 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-67ndf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-4bsxc_openshift-marketplace(bfb801bd-e785-4fdc-8a34-d2b8769d3a29): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 15:35:01 crc kubenswrapper[4894]: E1209 15:35:01.998489 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-4bsxc" podUID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" Dec 09 15:35:02 crc kubenswrapper[4894]: E1209 15:35:02.015703 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Dec 09 15:35:02 crc kubenswrapper[4894]: E1209 15:35:02.015883 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hwmhn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-q58q9_openshift-marketplace(ee374eee-998a-48f2-acdf-39daa5ce2fdd): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Dec 09 15:35:02 crc kubenswrapper[4894]: E1209 15:35:02.017303 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-q58q9" podUID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" Dec 09 15:35:02 crc kubenswrapper[4894]: I1209 15:35:02.321226 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Dec 09 15:35:02 crc kubenswrapper[4894]: I1209 15:35:02.324245 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Dec 09 15:35:02 crc kubenswrapper[4894]: I1209 15:35:02.405199 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-qxfr8"] Dec 09 15:35:02 crc kubenswrapper[4894]: W1209 15:35:02.412029 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod948b28c5_e35c_4e6c_b31c_c0d5b205389b.slice/crio-2b67e06e556a0d124e6010b3a00e9d6a66657eb800f509fb9eef95307ce1eb60 WatchSource:0}: Error finding container 2b67e06e556a0d124e6010b3a00e9d6a66657eb800f509fb9eef95307ce1eb60: Status 404 returned error can't find the container with id 2b67e06e556a0d124e6010b3a00e9d6a66657eb800f509fb9eef95307ce1eb60 Dec 09 15:35:02 crc kubenswrapper[4894]: I1209 15:35:02.698592 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-qxfr8" event={"ID":"948b28c5-e35c-4e6c-b31c-c0d5b205389b","Type":"ContainerStarted","Data":"a8c37b5a2afddd5e6d4fd442be55b32df17da0a8d9a6b05a9041b9b735812488"} Dec 09 15:35:02 crc kubenswrapper[4894]: I1209 15:35:02.698653 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-qxfr8" event={"ID":"948b28c5-e35c-4e6c-b31c-c0d5b205389b","Type":"ContainerStarted","Data":"2b67e06e556a0d124e6010b3a00e9d6a66657eb800f509fb9eef95307ce1eb60"} Dec 09 15:35:02 crc kubenswrapper[4894]: I1209 15:35:02.700935 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0a72fe2b-51af-49e9-a59f-de142120033b","Type":"ContainerStarted","Data":"1b65b8c8ff0c3e4ef99a385e13ee06b945b97973c7aeabf94066a99fd204b503"} Dec 09 15:35:02 crc kubenswrapper[4894]: I1209 15:35:02.701003 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0a72fe2b-51af-49e9-a59f-de142120033b","Type":"ContainerStarted","Data":"a2cf3ddd4108577cb4f99daf5a4c247d9263349cc74d718601995c8a6616f97f"} Dec 09 15:35:02 crc kubenswrapper[4894]: I1209 15:35:02.702366 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"5498bfbc-4945-4ffc-a7e3-db254f6cbabb","Type":"ContainerStarted","Data":"b904cfe76587b901b9288cc1d68c2365639ec91bc6f425f9c52cea547ab4fb70"} Dec 09 15:35:02 crc kubenswrapper[4894]: I1209 15:35:02.702440 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"5498bfbc-4945-4ffc-a7e3-db254f6cbabb","Type":"ContainerStarted","Data":"5c7fa54a93153214f66ec8c33febd144673a5eda810bdac86a802261a64bfb0d"} Dec 09 15:35:02 crc kubenswrapper[4894]: E1209 15:35:02.703967 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-nxlqt" podUID="2421584b-8a4b-4cb1-97c0-cb41d8ec16b2" Dec 09 15:35:02 crc kubenswrapper[4894]: E1209 15:35:02.704020 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-s4rbb" podUID="46d076e7-ca39-4337-bacc-ea3984f3bb5d" Dec 09 15:35:02 crc kubenswrapper[4894]: E1209 15:35:02.704894 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-q58q9" podUID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" Dec 09 15:35:02 crc kubenswrapper[4894]: E1209 15:35:02.705233 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-4bsxc" podUID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" Dec 09 15:35:02 crc kubenswrapper[4894]: I1209 15:35:02.716706 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=14.71668424 podStartE2EDuration="14.71668424s" podCreationTimestamp="2025-12-09 15:34:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:35:02.712607912 +0000 UTC m=+197.031818601" watchObservedRunningTime="2025-12-09 15:35:02.71668424 +0000 UTC m=+197.035894919" Dec 09 15:35:02 crc kubenswrapper[4894]: I1209 15:35:02.730694 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=9.73066771 podStartE2EDuration="9.73066771s" podCreationTimestamp="2025-12-09 15:34:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:35:02.730488336 +0000 UTC m=+197.049699005" watchObservedRunningTime="2025-12-09 15:35:02.73066771 +0000 UTC m=+197.049878379" Dec 09 15:35:03 crc kubenswrapper[4894]: I1209 15:35:03.710985 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-qxfr8" event={"ID":"948b28c5-e35c-4e6c-b31c-c0d5b205389b","Type":"ContainerStarted","Data":"0e8c9643be29eb0fae6ca630a22f8f8b4fa184a6810d66ef152b0a8b1fd286eb"} Dec 09 15:35:03 crc kubenswrapper[4894]: I1209 15:35:03.716054 4894 generic.go:334] "Generic (PLEG): container finished" podID="0a72fe2b-51af-49e9-a59f-de142120033b" containerID="1b65b8c8ff0c3e4ef99a385e13ee06b945b97973c7aeabf94066a99fd204b503" exitCode=0 Dec 09 15:35:03 crc kubenswrapper[4894]: I1209 15:35:03.716323 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0a72fe2b-51af-49e9-a59f-de142120033b","Type":"ContainerDied","Data":"1b65b8c8ff0c3e4ef99a385e13ee06b945b97973c7aeabf94066a99fd204b503"} Dec 09 15:35:03 crc kubenswrapper[4894]: I1209 15:35:03.729812 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-qxfr8" podStartSLOduration=179.729791739 podStartE2EDuration="2m59.729791739s" podCreationTimestamp="2025-12-09 15:32:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:35:03.729316837 +0000 UTC m=+198.048527506" watchObservedRunningTime="2025-12-09 15:35:03.729791739 +0000 UTC m=+198.049002408" Dec 09 15:35:04 crc kubenswrapper[4894]: I1209 15:35:04.966578 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 15:35:05 crc kubenswrapper[4894]: I1209 15:35:05.116933 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0a72fe2b-51af-49e9-a59f-de142120033b-kubelet-dir\") pod \"0a72fe2b-51af-49e9-a59f-de142120033b\" (UID: \"0a72fe2b-51af-49e9-a59f-de142120033b\") " Dec 09 15:35:05 crc kubenswrapper[4894]: I1209 15:35:05.117027 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a72fe2b-51af-49e9-a59f-de142120033b-kube-api-access\") pod \"0a72fe2b-51af-49e9-a59f-de142120033b\" (UID: \"0a72fe2b-51af-49e9-a59f-de142120033b\") " Dec 09 15:35:05 crc kubenswrapper[4894]: I1209 15:35:05.117070 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0a72fe2b-51af-49e9-a59f-de142120033b-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "0a72fe2b-51af-49e9-a59f-de142120033b" (UID: "0a72fe2b-51af-49e9-a59f-de142120033b"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:35:05 crc kubenswrapper[4894]: I1209 15:35:05.122350 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0a72fe2b-51af-49e9-a59f-de142120033b-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0a72fe2b-51af-49e9-a59f-de142120033b" (UID: "0a72fe2b-51af-49e9-a59f-de142120033b"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:35:05 crc kubenswrapper[4894]: I1209 15:35:05.219013 4894 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/0a72fe2b-51af-49e9-a59f-de142120033b-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 09 15:35:05 crc kubenswrapper[4894]: I1209 15:35:05.219089 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0a72fe2b-51af-49e9-a59f-de142120033b-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 15:35:05 crc kubenswrapper[4894]: I1209 15:35:05.729109 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"0a72fe2b-51af-49e9-a59f-de142120033b","Type":"ContainerDied","Data":"a2cf3ddd4108577cb4f99daf5a4c247d9263349cc74d718601995c8a6616f97f"} Dec 09 15:35:05 crc kubenswrapper[4894]: I1209 15:35:05.729393 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a2cf3ddd4108577cb4f99daf5a4c247d9263349cc74d718601995c8a6616f97f" Dec 09 15:35:05 crc kubenswrapper[4894]: I1209 15:35:05.729180 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Dec 09 15:35:12 crc kubenswrapper[4894]: I1209 15:35:12.257411 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:35:12 crc kubenswrapper[4894]: I1209 15:35:12.258021 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:35:12 crc kubenswrapper[4894]: I1209 15:35:12.258081 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 15:35:12 crc kubenswrapper[4894]: I1209 15:35:12.258774 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f"} pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 15:35:12 crc kubenswrapper[4894]: I1209 15:35:12.258900 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" containerID="cri-o://f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f" gracePeriod=600 Dec 09 15:35:12 crc kubenswrapper[4894]: I1209 15:35:12.768978 4894 generic.go:334] "Generic (PLEG): container finished" podID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerID="f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f" exitCode=0 Dec 09 15:35:12 crc kubenswrapper[4894]: I1209 15:35:12.769067 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerDied","Data":"f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f"} Dec 09 15:35:12 crc kubenswrapper[4894]: I1209 15:35:12.769405 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerStarted","Data":"cff2cb26767fe11bd813d02b6541ceb81064036337295e671ed7939ae0026edf"} Dec 09 15:35:12 crc kubenswrapper[4894]: I1209 15:35:12.771970 4894 generic.go:334] "Generic (PLEG): container finished" podID="22822cfe-7b3b-4b3a-ba4f-c19630ccccae" containerID="e879b9be8f12f3b950969a4478d63ce631717814658a5c92f55542c738a676c8" exitCode=0 Dec 09 15:35:12 crc kubenswrapper[4894]: I1209 15:35:12.772002 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lz9jb" event={"ID":"22822cfe-7b3b-4b3a-ba4f-c19630ccccae","Type":"ContainerDied","Data":"e879b9be8f12f3b950969a4478d63ce631717814658a5c92f55542c738a676c8"} Dec 09 15:35:13 crc kubenswrapper[4894]: I1209 15:35:13.778182 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lz9jb" event={"ID":"22822cfe-7b3b-4b3a-ba4f-c19630ccccae","Type":"ContainerStarted","Data":"9eb43d6eb02749525fc0b809b001d309d90513c9d96bf1cfb234aa3abd9f1eb0"} Dec 09 15:35:13 crc kubenswrapper[4894]: I1209 15:35:13.798416 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-lz9jb" podStartSLOduration=2.8060797060000002 podStartE2EDuration="59.798395813s" podCreationTimestamp="2025-12-09 15:34:14 +0000 UTC" firstStartedPulling="2025-12-09 15:34:16.202034303 +0000 UTC m=+150.521244972" lastFinishedPulling="2025-12-09 15:35:13.19435041 +0000 UTC m=+207.513561079" observedRunningTime="2025-12-09 15:35:13.798047864 +0000 UTC m=+208.117258533" watchObservedRunningTime="2025-12-09 15:35:13.798395813 +0000 UTC m=+208.117606492" Dec 09 15:35:14 crc kubenswrapper[4894]: I1209 15:35:14.786510 4894 generic.go:334] "Generic (PLEG): container finished" podID="c278157d-d806-4fa1-a597-076d95a286e2" containerID="3dbbd9400817dcec45a38afefbd879bcf664787acea95df58a4d7763fd2f2d5a" exitCode=0 Dec 09 15:35:14 crc kubenswrapper[4894]: I1209 15:35:14.786558 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ff286" event={"ID":"c278157d-d806-4fa1-a597-076d95a286e2","Type":"ContainerDied","Data":"3dbbd9400817dcec45a38afefbd879bcf664787acea95df58a4d7763fd2f2d5a"} Dec 09 15:35:15 crc kubenswrapper[4894]: I1209 15:35:15.244022 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-lz9jb" Dec 09 15:35:15 crc kubenswrapper[4894]: I1209 15:35:15.244334 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-lz9jb" Dec 09 15:35:15 crc kubenswrapper[4894]: I1209 15:35:15.803588 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ff286" event={"ID":"c278157d-d806-4fa1-a597-076d95a286e2","Type":"ContainerStarted","Data":"a15756c0576e9c5301de22221f66cec7e5bdb908775678f43f6efeeb8dac247e"} Dec 09 15:35:15 crc kubenswrapper[4894]: I1209 15:35:15.821608 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-ff286" podStartSLOduration=3.516773113 podStartE2EDuration="1m4.821593772s" podCreationTimestamp="2025-12-09 15:34:11 +0000 UTC" firstStartedPulling="2025-12-09 15:34:14.106453073 +0000 UTC m=+148.425663732" lastFinishedPulling="2025-12-09 15:35:15.411273722 +0000 UTC m=+209.730484391" observedRunningTime="2025-12-09 15:35:15.819671652 +0000 UTC m=+210.138882341" watchObservedRunningTime="2025-12-09 15:35:15.821593772 +0000 UTC m=+210.140804441" Dec 09 15:35:16 crc kubenswrapper[4894]: I1209 15:35:16.295962 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-lz9jb" podUID="22822cfe-7b3b-4b3a-ba4f-c19630ccccae" containerName="registry-server" probeResult="failure" output=< Dec 09 15:35:16 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 09 15:35:16 crc kubenswrapper[4894]: > Dec 09 15:35:16 crc kubenswrapper[4894]: I1209 15:35:16.817972 4894 generic.go:334] "Generic (PLEG): container finished" podID="e042f2bf-1069-42a4-aef2-72ec5da57c99" containerID="e6681c65d467f358b6de495d6cc7075d5b0ce2984f6822bf75540e7c1a3ac03f" exitCode=0 Dec 09 15:35:16 crc kubenswrapper[4894]: I1209 15:35:16.818035 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxq7l" event={"ID":"e042f2bf-1069-42a4-aef2-72ec5da57c99","Type":"ContainerDied","Data":"e6681c65d467f358b6de495d6cc7075d5b0ce2984f6822bf75540e7c1a3ac03f"} Dec 09 15:35:16 crc kubenswrapper[4894]: I1209 15:35:16.820177 4894 generic.go:334] "Generic (PLEG): container finished" podID="46d076e7-ca39-4337-bacc-ea3984f3bb5d" containerID="76eb568fcf65d2a2a15572c2fac1c74773fd46fe7e5af72e45d1f970d43fed5d" exitCode=0 Dec 09 15:35:16 crc kubenswrapper[4894]: I1209 15:35:16.820714 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s4rbb" event={"ID":"46d076e7-ca39-4337-bacc-ea3984f3bb5d","Type":"ContainerDied","Data":"76eb568fcf65d2a2a15572c2fac1c74773fd46fe7e5af72e45d1f970d43fed5d"} Dec 09 15:35:16 crc kubenswrapper[4894]: I1209 15:35:16.823132 4894 generic.go:334] "Generic (PLEG): container finished" podID="4397f08d-278a-4e2d-9664-e44161c96638" containerID="6226e3de7692ce8337bd03063725bc9bc03bae168a181c1cfb5e6022767f3cc1" exitCode=0 Dec 09 15:35:16 crc kubenswrapper[4894]: I1209 15:35:16.823207 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hq8h6" event={"ID":"4397f08d-278a-4e2d-9664-e44161c96638","Type":"ContainerDied","Data":"6226e3de7692ce8337bd03063725bc9bc03bae168a181c1cfb5e6022767f3cc1"} Dec 09 15:35:16 crc kubenswrapper[4894]: I1209 15:35:16.827424 4894 generic.go:334] "Generic (PLEG): container finished" podID="2421584b-8a4b-4cb1-97c0-cb41d8ec16b2" containerID="ca8726e6cc5f2a3db199c57f3943b724e01f4dd5bfcbe321417b335da1888b7a" exitCode=0 Dec 09 15:35:16 crc kubenswrapper[4894]: I1209 15:35:16.827501 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxlqt" event={"ID":"2421584b-8a4b-4cb1-97c0-cb41d8ec16b2","Type":"ContainerDied","Data":"ca8726e6cc5f2a3db199c57f3943b724e01f4dd5bfcbe321417b335da1888b7a"} Dec 09 15:35:16 crc kubenswrapper[4894]: I1209 15:35:16.830849 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bsxc" event={"ID":"bfb801bd-e785-4fdc-8a34-d2b8769d3a29","Type":"ContainerStarted","Data":"d36d2a68f74ea0b8f46707474565374cac5914f5a72ff6b285884a6db33d90ce"} Dec 09 15:35:18 crc kubenswrapper[4894]: I1209 15:35:18.846759 4894 generic.go:334] "Generic (PLEG): container finished" podID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" containerID="d36d2a68f74ea0b8f46707474565374cac5914f5a72ff6b285884a6db33d90ce" exitCode=0 Dec 09 15:35:18 crc kubenswrapper[4894]: I1209 15:35:18.847908 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bsxc" event={"ID":"bfb801bd-e785-4fdc-8a34-d2b8769d3a29","Type":"ContainerDied","Data":"d36d2a68f74ea0b8f46707474565374cac5914f5a72ff6b285884a6db33d90ce"} Dec 09 15:35:22 crc kubenswrapper[4894]: I1209 15:35:22.494693 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-ff286" Dec 09 15:35:22 crc kubenswrapper[4894]: I1209 15:35:22.495049 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-ff286" Dec 09 15:35:22 crc kubenswrapper[4894]: I1209 15:35:22.773955 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-ff286" Dec 09 15:35:22 crc kubenswrapper[4894]: I1209 15:35:22.873142 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxlqt" event={"ID":"2421584b-8a4b-4cb1-97c0-cb41d8ec16b2","Type":"ContainerStarted","Data":"0a2bf7f94932408830073b3ab94b03ba306028cef6f083fe1fa991aad52155fb"} Dec 09 15:35:22 crc kubenswrapper[4894]: I1209 15:35:22.912612 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-ff286" Dec 09 15:35:23 crc kubenswrapper[4894]: I1209 15:35:23.896260 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-nxlqt" podStartSLOduration=5.285958255 podStartE2EDuration="1m10.896239561s" podCreationTimestamp="2025-12-09 15:34:13 +0000 UTC" firstStartedPulling="2025-12-09 15:34:16.214110817 +0000 UTC m=+150.533321486" lastFinishedPulling="2025-12-09 15:35:21.824392123 +0000 UTC m=+216.143602792" observedRunningTime="2025-12-09 15:35:23.895208284 +0000 UTC m=+218.214418963" watchObservedRunningTime="2025-12-09 15:35:23.896239561 +0000 UTC m=+218.215450250" Dec 09 15:35:25 crc kubenswrapper[4894]: I1209 15:35:25.298476 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-lz9jb" Dec 09 15:35:25 crc kubenswrapper[4894]: I1209 15:35:25.343971 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-lz9jb" Dec 09 15:35:26 crc kubenswrapper[4894]: I1209 15:35:26.896899 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxq7l" event={"ID":"e042f2bf-1069-42a4-aef2-72ec5da57c99","Type":"ContainerStarted","Data":"f7d919ef7b3d5453ab9a54afd4239af89d4f62f1cd46dea9c88a8e314a42001a"} Dec 09 15:35:26 crc kubenswrapper[4894]: I1209 15:35:26.899183 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s4rbb" event={"ID":"46d076e7-ca39-4337-bacc-ea3984f3bb5d","Type":"ContainerStarted","Data":"a7d4f8a5a72a970c0079a043ac841d16ed17162bf97c5965d82a1c3d25a418d5"} Dec 09 15:35:26 crc kubenswrapper[4894]: I1209 15:35:26.902249 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hq8h6" event={"ID":"4397f08d-278a-4e2d-9664-e44161c96638","Type":"ContainerStarted","Data":"f99ea2eb94afcf5102569043baa964d088bfa60c0648c833f9019a560c201378"} Dec 09 15:35:26 crc kubenswrapper[4894]: I1209 15:35:26.904588 4894 generic.go:334] "Generic (PLEG): container finished" podID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" containerID="c695feb4cb7d0efbde1fd625bdfb372189c361bd0398971d0af9488f048f090a" exitCode=0 Dec 09 15:35:26 crc kubenswrapper[4894]: I1209 15:35:26.904627 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q58q9" event={"ID":"ee374eee-998a-48f2-acdf-39daa5ce2fdd","Type":"ContainerDied","Data":"c695feb4cb7d0efbde1fd625bdfb372189c361bd0398971d0af9488f048f090a"} Dec 09 15:35:26 crc kubenswrapper[4894]: I1209 15:35:26.918584 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-fxq7l" podStartSLOduration=4.602059015 podStartE2EDuration="1m15.91856381s" podCreationTimestamp="2025-12-09 15:34:11 +0000 UTC" firstStartedPulling="2025-12-09 15:34:14.106886454 +0000 UTC m=+148.426097113" lastFinishedPulling="2025-12-09 15:35:25.423391219 +0000 UTC m=+219.742601908" observedRunningTime="2025-12-09 15:35:26.917326457 +0000 UTC m=+221.236537146" watchObservedRunningTime="2025-12-09 15:35:26.91856381 +0000 UTC m=+221.237774479" Dec 09 15:35:26 crc kubenswrapper[4894]: I1209 15:35:26.940916 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-s4rbb" podStartSLOduration=4.644391325 podStartE2EDuration="1m15.940894322s" podCreationTimestamp="2025-12-09 15:34:11 +0000 UTC" firstStartedPulling="2025-12-09 15:34:14.164839538 +0000 UTC m=+148.484050207" lastFinishedPulling="2025-12-09 15:35:25.461342505 +0000 UTC m=+219.780553204" observedRunningTime="2025-12-09 15:35:26.93819076 +0000 UTC m=+221.257401439" watchObservedRunningTime="2025-12-09 15:35:26.940894322 +0000 UTC m=+221.260104991" Dec 09 15:35:26 crc kubenswrapper[4894]: I1209 15:35:26.993524 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hq8h6" podStartSLOduration=3.610731164 podStartE2EDuration="1m15.993507706s" podCreationTimestamp="2025-12-09 15:34:11 +0000 UTC" firstStartedPulling="2025-12-09 15:34:13.071950478 +0000 UTC m=+147.391161527" lastFinishedPulling="2025-12-09 15:35:25.4547274 +0000 UTC m=+219.773938069" observedRunningTime="2025-12-09 15:35:26.977678526 +0000 UTC m=+221.296889195" watchObservedRunningTime="2025-12-09 15:35:26.993507706 +0000 UTC m=+221.312718375" Dec 09 15:35:27 crc kubenswrapper[4894]: I1209 15:35:27.938161 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lz9jb"] Dec 09 15:35:27 crc kubenswrapper[4894]: I1209 15:35:27.939460 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-lz9jb" podUID="22822cfe-7b3b-4b3a-ba4f-c19630ccccae" containerName="registry-server" containerID="cri-o://9eb43d6eb02749525fc0b809b001d309d90513c9d96bf1cfb234aa3abd9f1eb0" gracePeriod=2 Dec 09 15:35:29 crc kubenswrapper[4894]: I1209 15:35:29.921584 4894 generic.go:334] "Generic (PLEG): container finished" podID="22822cfe-7b3b-4b3a-ba4f-c19630ccccae" containerID="9eb43d6eb02749525fc0b809b001d309d90513c9d96bf1cfb234aa3abd9f1eb0" exitCode=0 Dec 09 15:35:29 crc kubenswrapper[4894]: I1209 15:35:29.921623 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lz9jb" event={"ID":"22822cfe-7b3b-4b3a-ba4f-c19630ccccae","Type":"ContainerDied","Data":"9eb43d6eb02749525fc0b809b001d309d90513c9d96bf1cfb234aa3abd9f1eb0"} Dec 09 15:35:30 crc kubenswrapper[4894]: I1209 15:35:30.744817 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lz9jb" Dec 09 15:35:30 crc kubenswrapper[4894]: I1209 15:35:30.763497 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22822cfe-7b3b-4b3a-ba4f-c19630ccccae-utilities\") pod \"22822cfe-7b3b-4b3a-ba4f-c19630ccccae\" (UID: \"22822cfe-7b3b-4b3a-ba4f-c19630ccccae\") " Dec 09 15:35:30 crc kubenswrapper[4894]: I1209 15:35:30.763574 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gckhc\" (UniqueName: \"kubernetes.io/projected/22822cfe-7b3b-4b3a-ba4f-c19630ccccae-kube-api-access-gckhc\") pod \"22822cfe-7b3b-4b3a-ba4f-c19630ccccae\" (UID: \"22822cfe-7b3b-4b3a-ba4f-c19630ccccae\") " Dec 09 15:35:30 crc kubenswrapper[4894]: I1209 15:35:30.763609 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22822cfe-7b3b-4b3a-ba4f-c19630ccccae-catalog-content\") pod \"22822cfe-7b3b-4b3a-ba4f-c19630ccccae\" (UID: \"22822cfe-7b3b-4b3a-ba4f-c19630ccccae\") " Dec 09 15:35:30 crc kubenswrapper[4894]: I1209 15:35:30.765137 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22822cfe-7b3b-4b3a-ba4f-c19630ccccae-utilities" (OuterVolumeSpecName: "utilities") pod "22822cfe-7b3b-4b3a-ba4f-c19630ccccae" (UID: "22822cfe-7b3b-4b3a-ba4f-c19630ccccae"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:35:30 crc kubenswrapper[4894]: I1209 15:35:30.769415 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22822cfe-7b3b-4b3a-ba4f-c19630ccccae-kube-api-access-gckhc" (OuterVolumeSpecName: "kube-api-access-gckhc") pod "22822cfe-7b3b-4b3a-ba4f-c19630ccccae" (UID: "22822cfe-7b3b-4b3a-ba4f-c19630ccccae"). InnerVolumeSpecName "kube-api-access-gckhc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:35:30 crc kubenswrapper[4894]: I1209 15:35:30.866296 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/22822cfe-7b3b-4b3a-ba4f-c19630ccccae-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 15:35:30 crc kubenswrapper[4894]: I1209 15:35:30.866328 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gckhc\" (UniqueName: \"kubernetes.io/projected/22822cfe-7b3b-4b3a-ba4f-c19630ccccae-kube-api-access-gckhc\") on node \"crc\" DevicePath \"\"" Dec 09 15:35:30 crc kubenswrapper[4894]: I1209 15:35:30.873998 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/22822cfe-7b3b-4b3a-ba4f-c19630ccccae-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "22822cfe-7b3b-4b3a-ba4f-c19630ccccae" (UID: "22822cfe-7b3b-4b3a-ba4f-c19630ccccae"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:35:30 crc kubenswrapper[4894]: I1209 15:35:30.933661 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-lz9jb" event={"ID":"22822cfe-7b3b-4b3a-ba4f-c19630ccccae","Type":"ContainerDied","Data":"02255fc0367f25e7867f62c6f0b2589b76b9b10ba9d943abfad2db2644cee693"} Dec 09 15:35:30 crc kubenswrapper[4894]: I1209 15:35:30.933722 4894 scope.go:117] "RemoveContainer" containerID="9eb43d6eb02749525fc0b809b001d309d90513c9d96bf1cfb234aa3abd9f1eb0" Dec 09 15:35:30 crc kubenswrapper[4894]: I1209 15:35:30.933735 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-lz9jb" Dec 09 15:35:30 crc kubenswrapper[4894]: I1209 15:35:30.966665 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-lz9jb"] Dec 09 15:35:30 crc kubenswrapper[4894]: I1209 15:35:30.967063 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/22822cfe-7b3b-4b3a-ba4f-c19630ccccae-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 15:35:30 crc kubenswrapper[4894]: I1209 15:35:30.969498 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-lz9jb"] Dec 09 15:35:31 crc kubenswrapper[4894]: I1209 15:35:31.214228 4894 scope.go:117] "RemoveContainer" containerID="e879b9be8f12f3b950969a4478d63ce631717814658a5c92f55542c738a676c8" Dec 09 15:35:31 crc kubenswrapper[4894]: I1209 15:35:31.239408 4894 scope.go:117] "RemoveContainer" containerID="52c5871e23c7392d33bdc1375b7d6206b816af32b465ffcd85d034674e5c4893" Dec 09 15:35:31 crc kubenswrapper[4894]: I1209 15:35:31.644853 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hq8h6" Dec 09 15:35:31 crc kubenswrapper[4894]: I1209 15:35:31.645178 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hq8h6" Dec 09 15:35:31 crc kubenswrapper[4894]: I1209 15:35:31.691605 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hq8h6" Dec 09 15:35:31 crc kubenswrapper[4894]: I1209 15:35:31.980395 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hq8h6" Dec 09 15:35:32 crc kubenswrapper[4894]: I1209 15:35:32.032620 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-fxq7l" Dec 09 15:35:32 crc kubenswrapper[4894]: I1209 15:35:32.032694 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-fxq7l" Dec 09 15:35:32 crc kubenswrapper[4894]: I1209 15:35:32.069231 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-fxq7l" Dec 09 15:35:32 crc kubenswrapper[4894]: I1209 15:35:32.113609 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22822cfe-7b3b-4b3a-ba4f-c19630ccccae" path="/var/lib/kubelet/pods/22822cfe-7b3b-4b3a-ba4f-c19630ccccae/volumes" Dec 09 15:35:32 crc kubenswrapper[4894]: I1209 15:35:32.495107 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-s4rbb" Dec 09 15:35:32 crc kubenswrapper[4894]: I1209 15:35:32.495428 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-s4rbb" Dec 09 15:35:32 crc kubenswrapper[4894]: I1209 15:35:32.533372 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-s4rbb" Dec 09 15:35:32 crc kubenswrapper[4894]: I1209 15:35:32.947610 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bsxc" event={"ID":"bfb801bd-e785-4fdc-8a34-d2b8769d3a29","Type":"ContainerStarted","Data":"760697f2531f933eeec09e43003f58c82b1077c976a4af23b6f5ae395cdd8a8a"} Dec 09 15:35:32 crc kubenswrapper[4894]: I1209 15:35:32.991003 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-s4rbb" Dec 09 15:35:33 crc kubenswrapper[4894]: I1209 15:35:33.002988 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-fxq7l" Dec 09 15:35:33 crc kubenswrapper[4894]: I1209 15:35:33.636088 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-nxlqt" Dec 09 15:35:33 crc kubenswrapper[4894]: I1209 15:35:33.636158 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-nxlqt" Dec 09 15:35:33 crc kubenswrapper[4894]: I1209 15:35:33.674834 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-nxlqt" Dec 09 15:35:33 crc kubenswrapper[4894]: I1209 15:35:33.976274 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4bsxc" podStartSLOduration=4.943699721 podStartE2EDuration="1m19.976232466s" podCreationTimestamp="2025-12-09 15:34:14 +0000 UTC" firstStartedPulling="2025-12-09 15:34:16.181967895 +0000 UTC m=+150.501178564" lastFinishedPulling="2025-12-09 15:35:31.21450063 +0000 UTC m=+225.533711309" observedRunningTime="2025-12-09 15:35:33.9729871 +0000 UTC m=+228.292197809" watchObservedRunningTime="2025-12-09 15:35:33.976232466 +0000 UTC m=+228.295443145" Dec 09 15:35:34 crc kubenswrapper[4894]: I1209 15:35:34.003527 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-nxlqt" Dec 09 15:35:34 crc kubenswrapper[4894]: I1209 15:35:34.333110 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fxq7l"] Dec 09 15:35:34 crc kubenswrapper[4894]: I1209 15:35:34.846950 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4bsxc" Dec 09 15:35:34 crc kubenswrapper[4894]: I1209 15:35:34.846999 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4bsxc" Dec 09 15:35:34 crc kubenswrapper[4894]: I1209 15:35:34.959484 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q58q9" event={"ID":"ee374eee-998a-48f2-acdf-39daa5ce2fdd","Type":"ContainerStarted","Data":"6887e8e0a215db3554aceb2634a1e253366e560bf7d5cde02b24ee31510ac3bb"} Dec 09 15:35:34 crc kubenswrapper[4894]: I1209 15:35:34.959620 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-fxq7l" podUID="e042f2bf-1069-42a4-aef2-72ec5da57c99" containerName="registry-server" containerID="cri-o://f7d919ef7b3d5453ab9a54afd4239af89d4f62f1cd46dea9c88a8e314a42001a" gracePeriod=2 Dec 09 15:35:34 crc kubenswrapper[4894]: I1209 15:35:34.994919 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q58q9" podStartSLOduration=4.837153606 podStartE2EDuration="1m21.994898403s" podCreationTimestamp="2025-12-09 15:34:13 +0000 UTC" firstStartedPulling="2025-12-09 15:34:16.185367667 +0000 UTC m=+150.504578336" lastFinishedPulling="2025-12-09 15:35:33.343112464 +0000 UTC m=+227.662323133" observedRunningTime="2025-12-09 15:35:34.993979139 +0000 UTC m=+229.313189808" watchObservedRunningTime="2025-12-09 15:35:34.994898403 +0000 UTC m=+229.314109072" Dec 09 15:35:35 crc kubenswrapper[4894]: I1209 15:35:35.338931 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fxq7l" Dec 09 15:35:35 crc kubenswrapper[4894]: I1209 15:35:35.523078 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqpnd\" (UniqueName: \"kubernetes.io/projected/e042f2bf-1069-42a4-aef2-72ec5da57c99-kube-api-access-gqpnd\") pod \"e042f2bf-1069-42a4-aef2-72ec5da57c99\" (UID: \"e042f2bf-1069-42a4-aef2-72ec5da57c99\") " Dec 09 15:35:35 crc kubenswrapper[4894]: I1209 15:35:35.523175 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e042f2bf-1069-42a4-aef2-72ec5da57c99-utilities\") pod \"e042f2bf-1069-42a4-aef2-72ec5da57c99\" (UID: \"e042f2bf-1069-42a4-aef2-72ec5da57c99\") " Dec 09 15:35:35 crc kubenswrapper[4894]: I1209 15:35:35.523205 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e042f2bf-1069-42a4-aef2-72ec5da57c99-catalog-content\") pod \"e042f2bf-1069-42a4-aef2-72ec5da57c99\" (UID: \"e042f2bf-1069-42a4-aef2-72ec5da57c99\") " Dec 09 15:35:35 crc kubenswrapper[4894]: I1209 15:35:35.536270 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e042f2bf-1069-42a4-aef2-72ec5da57c99-utilities" (OuterVolumeSpecName: "utilities") pod "e042f2bf-1069-42a4-aef2-72ec5da57c99" (UID: "e042f2bf-1069-42a4-aef2-72ec5da57c99"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:35:35 crc kubenswrapper[4894]: I1209 15:35:35.541162 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e042f2bf-1069-42a4-aef2-72ec5da57c99-kube-api-access-gqpnd" (OuterVolumeSpecName: "kube-api-access-gqpnd") pod "e042f2bf-1069-42a4-aef2-72ec5da57c99" (UID: "e042f2bf-1069-42a4-aef2-72ec5da57c99"). InnerVolumeSpecName "kube-api-access-gqpnd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:35:35 crc kubenswrapper[4894]: I1209 15:35:35.568694 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e042f2bf-1069-42a4-aef2-72ec5da57c99-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e042f2bf-1069-42a4-aef2-72ec5da57c99" (UID: "e042f2bf-1069-42a4-aef2-72ec5da57c99"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:35:35 crc kubenswrapper[4894]: I1209 15:35:35.624041 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqpnd\" (UniqueName: \"kubernetes.io/projected/e042f2bf-1069-42a4-aef2-72ec5da57c99-kube-api-access-gqpnd\") on node \"crc\" DevicePath \"\"" Dec 09 15:35:35 crc kubenswrapper[4894]: I1209 15:35:35.624077 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e042f2bf-1069-42a4-aef2-72ec5da57c99-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 15:35:35 crc kubenswrapper[4894]: I1209 15:35:35.624088 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e042f2bf-1069-42a4-aef2-72ec5da57c99-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 15:35:35 crc kubenswrapper[4894]: I1209 15:35:35.894159 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4bsxc" podUID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" containerName="registry-server" probeResult="failure" output=< Dec 09 15:35:35 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 09 15:35:35 crc kubenswrapper[4894]: > Dec 09 15:35:35 crc kubenswrapper[4894]: I1209 15:35:35.966397 4894 generic.go:334] "Generic (PLEG): container finished" podID="e042f2bf-1069-42a4-aef2-72ec5da57c99" containerID="f7d919ef7b3d5453ab9a54afd4239af89d4f62f1cd46dea9c88a8e314a42001a" exitCode=0 Dec 09 15:35:35 crc kubenswrapper[4894]: I1209 15:35:35.966448 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-fxq7l" Dec 09 15:35:35 crc kubenswrapper[4894]: I1209 15:35:35.966448 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxq7l" event={"ID":"e042f2bf-1069-42a4-aef2-72ec5da57c99","Type":"ContainerDied","Data":"f7d919ef7b3d5453ab9a54afd4239af89d4f62f1cd46dea9c88a8e314a42001a"} Dec 09 15:35:35 crc kubenswrapper[4894]: I1209 15:35:35.966616 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-fxq7l" event={"ID":"e042f2bf-1069-42a4-aef2-72ec5da57c99","Type":"ContainerDied","Data":"f338cf6ba9baffe040fc08bc935fda57c2613b4df4d52db7f30c58c48997057f"} Dec 09 15:35:35 crc kubenswrapper[4894]: I1209 15:35:35.966667 4894 scope.go:117] "RemoveContainer" containerID="f7d919ef7b3d5453ab9a54afd4239af89d4f62f1cd46dea9c88a8e314a42001a" Dec 09 15:35:35 crc kubenswrapper[4894]: I1209 15:35:35.983628 4894 scope.go:117] "RemoveContainer" containerID="e6681c65d467f358b6de495d6cc7075d5b0ce2984f6822bf75540e7c1a3ac03f" Dec 09 15:35:35 crc kubenswrapper[4894]: I1209 15:35:35.996039 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-fxq7l"] Dec 09 15:35:35 crc kubenswrapper[4894]: I1209 15:35:35.996099 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-fxq7l"] Dec 09 15:35:36 crc kubenswrapper[4894]: I1209 15:35:36.015149 4894 scope.go:117] "RemoveContainer" containerID="5cb2340301f99f23c4d77ce927ea630a9c59e1adb619f5e4605d2d3f5a8a65cf" Dec 09 15:35:36 crc kubenswrapper[4894]: I1209 15:35:36.029333 4894 scope.go:117] "RemoveContainer" containerID="f7d919ef7b3d5453ab9a54afd4239af89d4f62f1cd46dea9c88a8e314a42001a" Dec 09 15:35:36 crc kubenswrapper[4894]: E1209 15:35:36.029785 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7d919ef7b3d5453ab9a54afd4239af89d4f62f1cd46dea9c88a8e314a42001a\": container with ID starting with f7d919ef7b3d5453ab9a54afd4239af89d4f62f1cd46dea9c88a8e314a42001a not found: ID does not exist" containerID="f7d919ef7b3d5453ab9a54afd4239af89d4f62f1cd46dea9c88a8e314a42001a" Dec 09 15:35:36 crc kubenswrapper[4894]: I1209 15:35:36.029834 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7d919ef7b3d5453ab9a54afd4239af89d4f62f1cd46dea9c88a8e314a42001a"} err="failed to get container status \"f7d919ef7b3d5453ab9a54afd4239af89d4f62f1cd46dea9c88a8e314a42001a\": rpc error: code = NotFound desc = could not find container \"f7d919ef7b3d5453ab9a54afd4239af89d4f62f1cd46dea9c88a8e314a42001a\": container with ID starting with f7d919ef7b3d5453ab9a54afd4239af89d4f62f1cd46dea9c88a8e314a42001a not found: ID does not exist" Dec 09 15:35:36 crc kubenswrapper[4894]: I1209 15:35:36.029869 4894 scope.go:117] "RemoveContainer" containerID="e6681c65d467f358b6de495d6cc7075d5b0ce2984f6822bf75540e7c1a3ac03f" Dec 09 15:35:36 crc kubenswrapper[4894]: E1209 15:35:36.030233 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6681c65d467f358b6de495d6cc7075d5b0ce2984f6822bf75540e7c1a3ac03f\": container with ID starting with e6681c65d467f358b6de495d6cc7075d5b0ce2984f6822bf75540e7c1a3ac03f not found: ID does not exist" containerID="e6681c65d467f358b6de495d6cc7075d5b0ce2984f6822bf75540e7c1a3ac03f" Dec 09 15:35:36 crc kubenswrapper[4894]: I1209 15:35:36.030261 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6681c65d467f358b6de495d6cc7075d5b0ce2984f6822bf75540e7c1a3ac03f"} err="failed to get container status \"e6681c65d467f358b6de495d6cc7075d5b0ce2984f6822bf75540e7c1a3ac03f\": rpc error: code = NotFound desc = could not find container \"e6681c65d467f358b6de495d6cc7075d5b0ce2984f6822bf75540e7c1a3ac03f\": container with ID starting with e6681c65d467f358b6de495d6cc7075d5b0ce2984f6822bf75540e7c1a3ac03f not found: ID does not exist" Dec 09 15:35:36 crc kubenswrapper[4894]: I1209 15:35:36.030279 4894 scope.go:117] "RemoveContainer" containerID="5cb2340301f99f23c4d77ce927ea630a9c59e1adb619f5e4605d2d3f5a8a65cf" Dec 09 15:35:36 crc kubenswrapper[4894]: E1209 15:35:36.030494 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5cb2340301f99f23c4d77ce927ea630a9c59e1adb619f5e4605d2d3f5a8a65cf\": container with ID starting with 5cb2340301f99f23c4d77ce927ea630a9c59e1adb619f5e4605d2d3f5a8a65cf not found: ID does not exist" containerID="5cb2340301f99f23c4d77ce927ea630a9c59e1adb619f5e4605d2d3f5a8a65cf" Dec 09 15:35:36 crc kubenswrapper[4894]: I1209 15:35:36.030520 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5cb2340301f99f23c4d77ce927ea630a9c59e1adb619f5e4605d2d3f5a8a65cf"} err="failed to get container status \"5cb2340301f99f23c4d77ce927ea630a9c59e1adb619f5e4605d2d3f5a8a65cf\": rpc error: code = NotFound desc = could not find container \"5cb2340301f99f23c4d77ce927ea630a9c59e1adb619f5e4605d2d3f5a8a65cf\": container with ID starting with 5cb2340301f99f23c4d77ce927ea630a9c59e1adb619f5e4605d2d3f5a8a65cf not found: ID does not exist" Dec 09 15:35:36 crc kubenswrapper[4894]: I1209 15:35:36.112328 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e042f2bf-1069-42a4-aef2-72ec5da57c99" path="/var/lib/kubelet/pods/e042f2bf-1069-42a4-aef2-72ec5da57c99/volumes" Dec 09 15:35:36 crc kubenswrapper[4894]: I1209 15:35:36.735215 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s4rbb"] Dec 09 15:35:36 crc kubenswrapper[4894]: I1209 15:35:36.735614 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-s4rbb" podUID="46d076e7-ca39-4337-bacc-ea3984f3bb5d" containerName="registry-server" containerID="cri-o://a7d4f8a5a72a970c0079a043ac841d16ed17162bf97c5965d82a1c3d25a418d5" gracePeriod=2 Dec 09 15:35:36 crc kubenswrapper[4894]: I1209 15:35:36.976556 4894 generic.go:334] "Generic (PLEG): container finished" podID="46d076e7-ca39-4337-bacc-ea3984f3bb5d" containerID="a7d4f8a5a72a970c0079a043ac841d16ed17162bf97c5965d82a1c3d25a418d5" exitCode=0 Dec 09 15:35:36 crc kubenswrapper[4894]: I1209 15:35:36.976667 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s4rbb" event={"ID":"46d076e7-ca39-4337-bacc-ea3984f3bb5d","Type":"ContainerDied","Data":"a7d4f8a5a72a970c0079a043ac841d16ed17162bf97c5965d82a1c3d25a418d5"} Dec 09 15:35:37 crc kubenswrapper[4894]: I1209 15:35:37.111385 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s4rbb" Dec 09 15:35:37 crc kubenswrapper[4894]: I1209 15:35:37.139401 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d076e7-ca39-4337-bacc-ea3984f3bb5d-utilities\") pod \"46d076e7-ca39-4337-bacc-ea3984f3bb5d\" (UID: \"46d076e7-ca39-4337-bacc-ea3984f3bb5d\") " Dec 09 15:35:37 crc kubenswrapper[4894]: I1209 15:35:37.139478 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d076e7-ca39-4337-bacc-ea3984f3bb5d-catalog-content\") pod \"46d076e7-ca39-4337-bacc-ea3984f3bb5d\" (UID: \"46d076e7-ca39-4337-bacc-ea3984f3bb5d\") " Dec 09 15:35:37 crc kubenswrapper[4894]: I1209 15:35:37.141318 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46d076e7-ca39-4337-bacc-ea3984f3bb5d-utilities" (OuterVolumeSpecName: "utilities") pod "46d076e7-ca39-4337-bacc-ea3984f3bb5d" (UID: "46d076e7-ca39-4337-bacc-ea3984f3bb5d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:35:37 crc kubenswrapper[4894]: I1209 15:35:37.200757 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/46d076e7-ca39-4337-bacc-ea3984f3bb5d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "46d076e7-ca39-4337-bacc-ea3984f3bb5d" (UID: "46d076e7-ca39-4337-bacc-ea3984f3bb5d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:35:37 crc kubenswrapper[4894]: I1209 15:35:37.240359 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-trr6z\" (UniqueName: \"kubernetes.io/projected/46d076e7-ca39-4337-bacc-ea3984f3bb5d-kube-api-access-trr6z\") pod \"46d076e7-ca39-4337-bacc-ea3984f3bb5d\" (UID: \"46d076e7-ca39-4337-bacc-ea3984f3bb5d\") " Dec 09 15:35:37 crc kubenswrapper[4894]: I1209 15:35:37.240685 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/46d076e7-ca39-4337-bacc-ea3984f3bb5d-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 15:35:37 crc kubenswrapper[4894]: I1209 15:35:37.240702 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/46d076e7-ca39-4337-bacc-ea3984f3bb5d-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 15:35:37 crc kubenswrapper[4894]: I1209 15:35:37.246900 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46d076e7-ca39-4337-bacc-ea3984f3bb5d-kube-api-access-trr6z" (OuterVolumeSpecName: "kube-api-access-trr6z") pod "46d076e7-ca39-4337-bacc-ea3984f3bb5d" (UID: "46d076e7-ca39-4337-bacc-ea3984f3bb5d"). InnerVolumeSpecName "kube-api-access-trr6z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:35:37 crc kubenswrapper[4894]: I1209 15:35:37.341824 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-trr6z\" (UniqueName: \"kubernetes.io/projected/46d076e7-ca39-4337-bacc-ea3984f3bb5d-kube-api-access-trr6z\") on node \"crc\" DevicePath \"\"" Dec 09 15:35:37 crc kubenswrapper[4894]: I1209 15:35:37.986037 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-s4rbb" event={"ID":"46d076e7-ca39-4337-bacc-ea3984f3bb5d","Type":"ContainerDied","Data":"dd15011d9c6afe0129e5e4b73e8a3501887648369064fb7f075281ca284a2c08"} Dec 09 15:35:37 crc kubenswrapper[4894]: I1209 15:35:37.986107 4894 scope.go:117] "RemoveContainer" containerID="a7d4f8a5a72a970c0079a043ac841d16ed17162bf97c5965d82a1c3d25a418d5" Dec 09 15:35:37 crc kubenswrapper[4894]: I1209 15:35:37.986130 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-s4rbb" Dec 09 15:35:38 crc kubenswrapper[4894]: I1209 15:35:38.016853 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-s4rbb"] Dec 09 15:35:38 crc kubenswrapper[4894]: I1209 15:35:38.023424 4894 scope.go:117] "RemoveContainer" containerID="76eb568fcf65d2a2a15572c2fac1c74773fd46fe7e5af72e45d1f970d43fed5d" Dec 09 15:35:38 crc kubenswrapper[4894]: I1209 15:35:38.035023 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-s4rbb"] Dec 09 15:35:38 crc kubenswrapper[4894]: I1209 15:35:38.050470 4894 scope.go:117] "RemoveContainer" containerID="814bfae2a1ab16cee44409ed752fbecd54395276985b4cda425a0567223ddff5" Dec 09 15:35:38 crc kubenswrapper[4894]: E1209 15:35:38.105238 4894 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46d076e7_ca39_4337_bacc_ea3984f3bb5d.slice/crio-dd15011d9c6afe0129e5e4b73e8a3501887648369064fb7f075281ca284a2c08\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46d076e7_ca39_4337_bacc_ea3984f3bb5d.slice\": RecentStats: unable to find data in memory cache]" Dec 09 15:35:38 crc kubenswrapper[4894]: I1209 15:35:38.114665 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46d076e7-ca39-4337-bacc-ea3984f3bb5d" path="/var/lib/kubelet/pods/46d076e7-ca39-4337-bacc-ea3984f3bb5d/volumes" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.223528 4894 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 09 15:35:40 crc kubenswrapper[4894]: E1209 15:35:40.224069 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46d076e7-ca39-4337-bacc-ea3984f3bb5d" containerName="extract-utilities" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.224082 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="46d076e7-ca39-4337-bacc-ea3984f3bb5d" containerName="extract-utilities" Dec 09 15:35:40 crc kubenswrapper[4894]: E1209 15:35:40.224094 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46d076e7-ca39-4337-bacc-ea3984f3bb5d" containerName="registry-server" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.224100 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="46d076e7-ca39-4337-bacc-ea3984f3bb5d" containerName="registry-server" Dec 09 15:35:40 crc kubenswrapper[4894]: E1209 15:35:40.224109 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e042f2bf-1069-42a4-aef2-72ec5da57c99" containerName="registry-server" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.224116 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e042f2bf-1069-42a4-aef2-72ec5da57c99" containerName="registry-server" Dec 09 15:35:40 crc kubenswrapper[4894]: E1209 15:35:40.224130 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22822cfe-7b3b-4b3a-ba4f-c19630ccccae" containerName="extract-utilities" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.224135 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="22822cfe-7b3b-4b3a-ba4f-c19630ccccae" containerName="extract-utilities" Dec 09 15:35:40 crc kubenswrapper[4894]: E1209 15:35:40.224143 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46d076e7-ca39-4337-bacc-ea3984f3bb5d" containerName="extract-content" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.224148 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="46d076e7-ca39-4337-bacc-ea3984f3bb5d" containerName="extract-content" Dec 09 15:35:40 crc kubenswrapper[4894]: E1209 15:35:40.224157 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0a72fe2b-51af-49e9-a59f-de142120033b" containerName="pruner" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.224163 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0a72fe2b-51af-49e9-a59f-de142120033b" containerName="pruner" Dec 09 15:35:40 crc kubenswrapper[4894]: E1209 15:35:40.224171 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e042f2bf-1069-42a4-aef2-72ec5da57c99" containerName="extract-utilities" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.224176 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e042f2bf-1069-42a4-aef2-72ec5da57c99" containerName="extract-utilities" Dec 09 15:35:40 crc kubenswrapper[4894]: E1209 15:35:40.224185 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22822cfe-7b3b-4b3a-ba4f-c19630ccccae" containerName="extract-content" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.224190 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="22822cfe-7b3b-4b3a-ba4f-c19630ccccae" containerName="extract-content" Dec 09 15:35:40 crc kubenswrapper[4894]: E1209 15:35:40.224197 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22822cfe-7b3b-4b3a-ba4f-c19630ccccae" containerName="registry-server" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.224202 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="22822cfe-7b3b-4b3a-ba4f-c19630ccccae" containerName="registry-server" Dec 09 15:35:40 crc kubenswrapper[4894]: E1209 15:35:40.224208 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e042f2bf-1069-42a4-aef2-72ec5da57c99" containerName="extract-content" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.224214 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e042f2bf-1069-42a4-aef2-72ec5da57c99" containerName="extract-content" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.224320 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e042f2bf-1069-42a4-aef2-72ec5da57c99" containerName="registry-server" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.224328 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="46d076e7-ca39-4337-bacc-ea3984f3bb5d" containerName="registry-server" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.224335 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0a72fe2b-51af-49e9-a59f-de142120033b" containerName="pruner" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.224343 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="22822cfe-7b3b-4b3a-ba4f-c19630ccccae" containerName="registry-server" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.224617 4894 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.224725 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.224877 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca" gracePeriod=15 Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.224914 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba" gracePeriod=15 Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.225003 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2" gracePeriod=15 Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.225037 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10" gracePeriod=15 Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.225085 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295" gracePeriod=15 Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.225652 4894 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 09 15:35:40 crc kubenswrapper[4894]: E1209 15:35:40.225799 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.225807 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 15:35:40 crc kubenswrapper[4894]: E1209 15:35:40.225818 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.225824 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 09 15:35:40 crc kubenswrapper[4894]: E1209 15:35:40.225834 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.225841 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Dec 09 15:35:40 crc kubenswrapper[4894]: E1209 15:35:40.225852 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.225858 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 09 15:35:40 crc kubenswrapper[4894]: E1209 15:35:40.225864 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.225869 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 09 15:35:40 crc kubenswrapper[4894]: E1209 15:35:40.225877 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.225884 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.225971 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.225981 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.225987 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.225997 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.226006 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Dec 09 15:35:40 crc kubenswrapper[4894]: E1209 15:35:40.226086 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.226092 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.227259 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.260700 4894 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Readiness probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]log ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]api-openshift-apiserver-available ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]api-openshift-oauth-apiserver-available ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]informer-sync ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/generic-apiserver-start-informers ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/priority-and-fairness-filter ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/start-apiextensions-informers ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/start-apiextensions-controllers ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/crd-informer-synced ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/start-system-namespaces-controller ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/rbac/bootstrap-roles ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/bootstrap-controller ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/start-kube-aggregator-informers ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/apiservice-registration-controller ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/apiservice-discovery-controller ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]autoregister-completion ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/apiservice-openapi-controller ok Dec 09 15:35:40 crc kubenswrapper[4894]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 09 15:35:40 crc kubenswrapper[4894]: [-]shutdown failed: reason withheld Dec 09 15:35:40 crc kubenswrapper[4894]: readyz check failed Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.260760 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:35:40 crc kubenswrapper[4894]: E1209 15:35:40.286783 4894 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.64:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.384718 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.385005 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.385115 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.385226 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.385358 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.385467 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.385554 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.385667 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.486659 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.486725 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.486761 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.486799 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.486800 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.486832 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.486857 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.486883 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.486911 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.486925 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.486995 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.487033 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.487075 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.487127 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.487158 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.487287 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: I1209 15:35:40.588079 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 15:35:40 crc kubenswrapper[4894]: E1209 15:35:40.612580 4894 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.64:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187f9608211870bd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-09 15:35:40.611678397 +0000 UTC m=+234.930889066,LastTimestamp:2025-12-09 15:35:40.611678397 +0000 UTC m=+234.930889066,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 09 15:35:41 crc kubenswrapper[4894]: I1209 15:35:41.001929 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"52938b5900a58c22b64437e82dfa1fe2879e75dc1588385b0a6ddc308a841fa0"} Dec 09 15:35:41 crc kubenswrapper[4894]: E1209 15:35:41.560085 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:35:41Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:35:41Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:35:41Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:35:41Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:051f12c2ceb6780475c8b29895666cdb2bb8f3c351295854dc8f7949d3c8e977\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:ff56860406975b258e8f5e8d3150aa1e107dd9194685995bdfd24b0f76f9da85\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1625196363},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1b04dc5280afec7ca3cd7829e5b05e16ef2e579219e3d265ab21e3567dcde88c\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a911a6cd3919c14246902db9c21dfe7bb023caf184b310371d1dbef02cf0e030\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1209810253},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:a2149679a3b0827822ac22129efa300ab25d90a557f8961dbcfdda3f55222fca\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:f52ab4781f6f00120bb4d7222688ce6f64d4ae75218b6cc36cbf9a14ad4d4060\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1201829714},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1b1026c62413fa239fa4ff6541fe8bda656c1281867ad6ee2c848feccb13c97e\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:2b633ebdc901d19290af4dc2d09e2b59c504c0fc15a3fba410b0ce098e2d5753\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1141987142},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:41 crc kubenswrapper[4894]: E1209 15:35:41.560507 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:41 crc kubenswrapper[4894]: E1209 15:35:41.560886 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:41 crc kubenswrapper[4894]: E1209 15:35:41.561065 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:41 crc kubenswrapper[4894]: E1209 15:35:41.561224 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:41 crc kubenswrapper[4894]: E1209 15:35:41.561235 4894 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 15:35:41 crc kubenswrapper[4894]: E1209 15:35:41.798897 4894 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:41 crc kubenswrapper[4894]: E1209 15:35:41.799564 4894 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:41 crc kubenswrapper[4894]: E1209 15:35:41.799997 4894 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:41 crc kubenswrapper[4894]: E1209 15:35:41.800253 4894 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:41 crc kubenswrapper[4894]: E1209 15:35:41.800496 4894 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:41 crc kubenswrapper[4894]: I1209 15:35:41.800526 4894 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Dec 09 15:35:41 crc kubenswrapper[4894]: E1209 15:35:41.800799 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" interval="200ms" Dec 09 15:35:42 crc kubenswrapper[4894]: E1209 15:35:42.001530 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" interval="400ms" Dec 09 15:35:42 crc kubenswrapper[4894]: I1209 15:35:42.008499 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"051cca6ef214542ae81693b1c1177e599eba5687a444956d22beb2a398b4bbe3"} Dec 09 15:35:42 crc kubenswrapper[4894]: E1209 15:35:42.009172 4894 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.64:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 15:35:42 crc kubenswrapper[4894]: I1209 15:35:42.009276 4894 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:42 crc kubenswrapper[4894]: I1209 15:35:42.009994 4894 generic.go:334] "Generic (PLEG): container finished" podID="5498bfbc-4945-4ffc-a7e3-db254f6cbabb" containerID="b904cfe76587b901b9288cc1d68c2365639ec91bc6f425f9c52cea547ab4fb70" exitCode=0 Dec 09 15:35:42 crc kubenswrapper[4894]: I1209 15:35:42.010069 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"5498bfbc-4945-4ffc-a7e3-db254f6cbabb","Type":"ContainerDied","Data":"b904cfe76587b901b9288cc1d68c2365639ec91bc6f425f9c52cea547ab4fb70"} Dec 09 15:35:42 crc kubenswrapper[4894]: I1209 15:35:42.010472 4894 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:42 crc kubenswrapper[4894]: I1209 15:35:42.010761 4894 status_manager.go:851] "Failed to get status for pod" podUID="5498bfbc-4945-4ffc-a7e3-db254f6cbabb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:42 crc kubenswrapper[4894]: I1209 15:35:42.012073 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Dec 09 15:35:42 crc kubenswrapper[4894]: I1209 15:35:42.013207 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 09 15:35:42 crc kubenswrapper[4894]: I1209 15:35:42.013828 4894 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba" exitCode=0 Dec 09 15:35:42 crc kubenswrapper[4894]: I1209 15:35:42.013848 4894 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2" exitCode=0 Dec 09 15:35:42 crc kubenswrapper[4894]: I1209 15:35:42.013855 4894 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295" exitCode=0 Dec 09 15:35:42 crc kubenswrapper[4894]: I1209 15:35:42.013862 4894 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10" exitCode=2 Dec 09 15:35:42 crc kubenswrapper[4894]: I1209 15:35:42.014130 4894 scope.go:117] "RemoveContainer" containerID="5c6e9e79dfc4fb80e9f8a591291682217de6c03a344f27ffbe1dbc5784e77edd" Dec 09 15:35:42 crc kubenswrapper[4894]: E1209 15:35:42.402803 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" interval="800ms" Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.021188 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 09 15:35:43 crc kubenswrapper[4894]: E1209 15:35:43.025247 4894 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.64:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.163527 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.164225 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.164769 4894 status_manager.go:851] "Failed to get status for pod" podUID="5498bfbc-4945-4ffc-a7e3-db254f6cbabb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.165151 4894 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:43 crc kubenswrapper[4894]: E1209 15:35:43.204751 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" interval="1.6s" Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.254680 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.255231 4894 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.255526 4894 status_manager.go:851] "Failed to get status for pod" podUID="5498bfbc-4945-4ffc-a7e3-db254f6cbabb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.323319 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.323366 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.323427 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.323444 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.323515 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.323557 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.323742 4894 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.323755 4894 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.323763 4894 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.424875 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/5498bfbc-4945-4ffc-a7e3-db254f6cbabb-var-lock\") pod \"5498bfbc-4945-4ffc-a7e3-db254f6cbabb\" (UID: \"5498bfbc-4945-4ffc-a7e3-db254f6cbabb\") " Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.425260 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5498bfbc-4945-4ffc-a7e3-db254f6cbabb-kube-api-access\") pod \"5498bfbc-4945-4ffc-a7e3-db254f6cbabb\" (UID: \"5498bfbc-4945-4ffc-a7e3-db254f6cbabb\") " Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.424969 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5498bfbc-4945-4ffc-a7e3-db254f6cbabb-var-lock" (OuterVolumeSpecName: "var-lock") pod "5498bfbc-4945-4ffc-a7e3-db254f6cbabb" (UID: "5498bfbc-4945-4ffc-a7e3-db254f6cbabb"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.425348 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5498bfbc-4945-4ffc-a7e3-db254f6cbabb-kubelet-dir\") pod \"5498bfbc-4945-4ffc-a7e3-db254f6cbabb\" (UID: \"5498bfbc-4945-4ffc-a7e3-db254f6cbabb\") " Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.425414 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5498bfbc-4945-4ffc-a7e3-db254f6cbabb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "5498bfbc-4945-4ffc-a7e3-db254f6cbabb" (UID: "5498bfbc-4945-4ffc-a7e3-db254f6cbabb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.425619 4894 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/5498bfbc-4945-4ffc-a7e3-db254f6cbabb-kubelet-dir\") on node \"crc\" DevicePath \"\"" Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.425662 4894 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/5498bfbc-4945-4ffc-a7e3-db254f6cbabb-var-lock\") on node \"crc\" DevicePath \"\"" Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.430208 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5498bfbc-4945-4ffc-a7e3-db254f6cbabb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "5498bfbc-4945-4ffc-a7e3-db254f6cbabb" (UID: "5498bfbc-4945-4ffc-a7e3-db254f6cbabb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:35:43 crc kubenswrapper[4894]: I1209 15:35:43.526279 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/5498bfbc-4945-4ffc-a7e3-db254f6cbabb-kube-api-access\") on node \"crc\" DevicePath \"\"" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.033822 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q58q9" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.039217 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q58q9" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.054125 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.055428 4894 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca" exitCode=0 Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.055600 4894 scope.go:117] "RemoveContainer" containerID="c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.055877 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.058553 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"5498bfbc-4945-4ffc-a7e3-db254f6cbabb","Type":"ContainerDied","Data":"5c7fa54a93153214f66ec8c33febd144673a5eda810bdac86a802261a64bfb0d"} Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.058591 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5c7fa54a93153214f66ec8c33febd144673a5eda810bdac86a802261a64bfb0d" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.058774 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.072941 4894 scope.go:117] "RemoveContainer" containerID="baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.076023 4894 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.076245 4894 status_manager.go:851] "Failed to get status for pod" podUID="5498bfbc-4945-4ffc-a7e3-db254f6cbabb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.076479 4894 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.076682 4894 status_manager.go:851] "Failed to get status for pod" podUID="5498bfbc-4945-4ffc-a7e3-db254f6cbabb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.081953 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q58q9" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.082694 4894 status_manager.go:851] "Failed to get status for pod" podUID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" pod="openshift-marketplace/redhat-marketplace-q58q9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-q58q9\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.083095 4894 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.083322 4894 status_manager.go:851] "Failed to get status for pod" podUID="5498bfbc-4945-4ffc-a7e3-db254f6cbabb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.093041 4894 scope.go:117] "RemoveContainer" containerID="cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.104043 4894 scope.go:117] "RemoveContainer" containerID="f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.111930 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.116346 4894 scope.go:117] "RemoveContainer" containerID="6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.133324 4894 scope.go:117] "RemoveContainer" containerID="af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.156454 4894 scope.go:117] "RemoveContainer" containerID="c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba" Dec 09 15:35:44 crc kubenswrapper[4894]: E1209 15:35:44.157232 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\": container with ID starting with c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba not found: ID does not exist" containerID="c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.157293 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba"} err="failed to get container status \"c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\": rpc error: code = NotFound desc = could not find container \"c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba\": container with ID starting with c55fe0927400be13492372535f67c314690dc1a0b1d4191bd396e715975352ba not found: ID does not exist" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.157328 4894 scope.go:117] "RemoveContainer" containerID="baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2" Dec 09 15:35:44 crc kubenswrapper[4894]: E1209 15:35:44.158105 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\": container with ID starting with baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2 not found: ID does not exist" containerID="baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.158186 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2"} err="failed to get container status \"baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\": rpc error: code = NotFound desc = could not find container \"baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2\": container with ID starting with baa10db7af9dad1cbd3a36d6082a626d66c87e47d27272219041f602c684bea2 not found: ID does not exist" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.158239 4894 scope.go:117] "RemoveContainer" containerID="cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295" Dec 09 15:35:44 crc kubenswrapper[4894]: E1209 15:35:44.158667 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\": container with ID starting with cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295 not found: ID does not exist" containerID="cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.158732 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295"} err="failed to get container status \"cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\": rpc error: code = NotFound desc = could not find container \"cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295\": container with ID starting with cea68e510df812a673dbe64926eafd917142b8f300d6b24d4c82aa583aaf1295 not found: ID does not exist" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.158756 4894 scope.go:117] "RemoveContainer" containerID="f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10" Dec 09 15:35:44 crc kubenswrapper[4894]: E1209 15:35:44.159159 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\": container with ID starting with f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10 not found: ID does not exist" containerID="f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.159248 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10"} err="failed to get container status \"f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\": rpc error: code = NotFound desc = could not find container \"f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10\": container with ID starting with f50295093a58356f79d120f8b8a0a42edc6b15a34c81ae81a031105aeb77be10 not found: ID does not exist" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.159289 4894 scope.go:117] "RemoveContainer" containerID="6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca" Dec 09 15:35:44 crc kubenswrapper[4894]: E1209 15:35:44.159595 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\": container with ID starting with 6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca not found: ID does not exist" containerID="6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.159625 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca"} err="failed to get container status \"6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\": rpc error: code = NotFound desc = could not find container \"6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca\": container with ID starting with 6312624d07100075c7ed43bb815c006a02fb120145a637726e03ac388503c2ca not found: ID does not exist" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.159661 4894 scope.go:117] "RemoveContainer" containerID="af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6" Dec 09 15:35:44 crc kubenswrapper[4894]: E1209 15:35:44.159886 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\": container with ID starting with af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6 not found: ID does not exist" containerID="af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.159917 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6"} err="failed to get container status \"af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\": rpc error: code = NotFound desc = could not find container \"af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6\": container with ID starting with af42b11a35c0942eb57e349f7f470e2542b0eed5f419274f30ca386d11d4edb6 not found: ID does not exist" Dec 09 15:35:44 crc kubenswrapper[4894]: E1209 15:35:44.806368 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" interval="3.2s" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.883353 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4bsxc" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.883856 4894 status_manager.go:851] "Failed to get status for pod" podUID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" pod="openshift-marketplace/redhat-operators-4bsxc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-4bsxc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.884250 4894 status_manager.go:851] "Failed to get status for pod" podUID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" pod="openshift-marketplace/redhat-marketplace-q58q9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-q58q9\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.884685 4894 status_manager.go:851] "Failed to get status for pod" podUID="5498bfbc-4945-4ffc-a7e3-db254f6cbabb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.924089 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4bsxc" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.924591 4894 status_manager.go:851] "Failed to get status for pod" podUID="5498bfbc-4945-4ffc-a7e3-db254f6cbabb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.924952 4894 status_manager.go:851] "Failed to get status for pod" podUID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" pod="openshift-marketplace/redhat-operators-4bsxc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-4bsxc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:44 crc kubenswrapper[4894]: I1209 15:35:44.925464 4894 status_manager.go:851] "Failed to get status for pod" podUID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" pod="openshift-marketplace/redhat-marketplace-q58q9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-q58q9\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:45 crc kubenswrapper[4894]: I1209 15:35:45.105513 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q58q9" Dec 09 15:35:45 crc kubenswrapper[4894]: I1209 15:35:45.106179 4894 status_manager.go:851] "Failed to get status for pod" podUID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" pod="openshift-marketplace/redhat-operators-4bsxc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-4bsxc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:45 crc kubenswrapper[4894]: I1209 15:35:45.106720 4894 status_manager.go:851] "Failed to get status for pod" podUID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" pod="openshift-marketplace/redhat-marketplace-q58q9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-q58q9\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:45 crc kubenswrapper[4894]: I1209 15:35:45.106982 4894 status_manager.go:851] "Failed to get status for pod" podUID="5498bfbc-4945-4ffc-a7e3-db254f6cbabb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:46 crc kubenswrapper[4894]: I1209 15:35:46.110411 4894 status_manager.go:851] "Failed to get status for pod" podUID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" pod="openshift-marketplace/redhat-marketplace-q58q9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-q58q9\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:46 crc kubenswrapper[4894]: I1209 15:35:46.111436 4894 status_manager.go:851] "Failed to get status for pod" podUID="5498bfbc-4945-4ffc-a7e3-db254f6cbabb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:46 crc kubenswrapper[4894]: I1209 15:35:46.111891 4894 status_manager.go:851] "Failed to get status for pod" podUID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" pod="openshift-marketplace/redhat-operators-4bsxc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-4bsxc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:48 crc kubenswrapper[4894]: E1209 15:35:48.007743 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" interval="6.4s" Dec 09 15:35:50 crc kubenswrapper[4894]: E1209 15:35:50.034340 4894 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.64:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.187f9608211870bd openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-12-09 15:35:40.611678397 +0000 UTC m=+234.930889066,LastTimestamp:2025-12-09 15:35:40.611678397 +0000 UTC m=+234.930889066,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Dec 09 15:35:51 crc kubenswrapper[4894]: E1209 15:35:51.834141 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:35:51Z\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:35:51Z\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:35:51Z\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-12-09T15:35:51Z\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:051f12c2ceb6780475c8b29895666cdb2bb8f3c351295854dc8f7949d3c8e977\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:ff56860406975b258e8f5e8d3150aa1e107dd9194685995bdfd24b0f76f9da85\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1625196363},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1b04dc5280afec7ca3cd7829e5b05e16ef2e579219e3d265ab21e3567dcde88c\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:a911a6cd3919c14246902db9c21dfe7bb023caf184b310371d1dbef02cf0e030\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1209810253},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:a2149679a3b0827822ac22129efa300ab25d90a557f8961dbcfdda3f55222fca\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:f52ab4781f6f00120bb4d7222688ce6f64d4ae75218b6cc36cbf9a14ad4d4060\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1201829714},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1b1026c62413fa239fa4ff6541fe8bda656c1281867ad6ee2c848feccb13c97e\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:2b633ebdc901d19290af4dc2d09e2b59c504c0fc15a3fba410b0ce098e2d5753\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1141987142},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792}]}}\" for node \"crc\": Patch \"https://api-int.crc.testing:6443/api/v1/nodes/crc/status?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:51 crc kubenswrapper[4894]: E1209 15:35:51.835952 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:51 crc kubenswrapper[4894]: E1209 15:35:51.836254 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:51 crc kubenswrapper[4894]: E1209 15:35:51.836788 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:51 crc kubenswrapper[4894]: E1209 15:35:51.837074 4894 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": Get \"https://api-int.crc.testing:6443/api/v1/nodes/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:51 crc kubenswrapper[4894]: E1209 15:35:51.837094 4894 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Dec 09 15:35:52 crc kubenswrapper[4894]: E1209 15:35:52.149109 4894 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.64:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" volumeName="registry-storage" Dec 09 15:35:53 crc kubenswrapper[4894]: I1209 15:35:53.106425 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:35:53 crc kubenswrapper[4894]: I1209 15:35:53.108798 4894 status_manager.go:851] "Failed to get status for pod" podUID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" pod="openshift-marketplace/redhat-operators-4bsxc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-4bsxc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:53 crc kubenswrapper[4894]: I1209 15:35:53.109286 4894 status_manager.go:851] "Failed to get status for pod" podUID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" pod="openshift-marketplace/redhat-marketplace-q58q9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-q58q9\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:53 crc kubenswrapper[4894]: I1209 15:35:53.109485 4894 status_manager.go:851] "Failed to get status for pod" podUID="5498bfbc-4945-4ffc-a7e3-db254f6cbabb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:53 crc kubenswrapper[4894]: I1209 15:35:53.120566 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 09 15:35:53 crc kubenswrapper[4894]: I1209 15:35:53.120669 4894 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755" exitCode=1 Dec 09 15:35:53 crc kubenswrapper[4894]: I1209 15:35:53.120711 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755"} Dec 09 15:35:53 crc kubenswrapper[4894]: I1209 15:35:53.121346 4894 scope.go:117] "RemoveContainer" containerID="f38a12c5fe61f7afa4ba67cc65bbb8d53bb2f9f6b51783c2cab802607974a755" Dec 09 15:35:53 crc kubenswrapper[4894]: I1209 15:35:53.121595 4894 status_manager.go:851] "Failed to get status for pod" podUID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" pod="openshift-marketplace/redhat-marketplace-q58q9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-q58q9\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:53 crc kubenswrapper[4894]: I1209 15:35:53.121892 4894 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:53 crc kubenswrapper[4894]: I1209 15:35:53.122173 4894 status_manager.go:851] "Failed to get status for pod" podUID="5498bfbc-4945-4ffc-a7e3-db254f6cbabb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:53 crc kubenswrapper[4894]: I1209 15:35:53.122425 4894 status_manager.go:851] "Failed to get status for pod" podUID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" pod="openshift-marketplace/redhat-operators-4bsxc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-4bsxc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:53 crc kubenswrapper[4894]: I1209 15:35:53.125802 4894 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9b8358d3-529c-43c9-964c-fe3bae20f1f8" Dec 09 15:35:53 crc kubenswrapper[4894]: I1209 15:35:53.125861 4894 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9b8358d3-529c-43c9-964c-fe3bae20f1f8" Dec 09 15:35:53 crc kubenswrapper[4894]: E1209 15:35:53.126275 4894 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:35:53 crc kubenswrapper[4894]: I1209 15:35:53.126857 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:35:53 crc kubenswrapper[4894]: W1209 15:35:53.157452 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-38db1336b60989964b2489b8ff0738d026284818a07163060c4912369632c736 WatchSource:0}: Error finding container 38db1336b60989964b2489b8ff0738d026284818a07163060c4912369632c736: Status 404 returned error can't find the container with id 38db1336b60989964b2489b8ff0738d026284818a07163060c4912369632c736 Dec 09 15:35:54 crc kubenswrapper[4894]: I1209 15:35:54.125197 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"38db1336b60989964b2489b8ff0738d026284818a07163060c4912369632c736"} Dec 09 15:35:54 crc kubenswrapper[4894]: E1209 15:35:54.409512 4894 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.64:6443: connect: connection refused" interval="7s" Dec 09 15:35:55 crc kubenswrapper[4894]: I1209 15:35:55.025038 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 15:35:55 crc kubenswrapper[4894]: I1209 15:35:55.133989 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Dec 09 15:35:55 crc kubenswrapper[4894]: I1209 15:35:55.134905 4894 status_manager.go:851] "Failed to get status for pod" podUID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" pod="openshift-marketplace/redhat-marketplace-q58q9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-q58q9\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:55 crc kubenswrapper[4894]: I1209 15:35:55.135024 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"147ddc96093f913503c0d5d6fa6bee345dcc6b3efeab8dbdda4cf28b80a22fe9"} Dec 09 15:35:55 crc kubenswrapper[4894]: I1209 15:35:55.135190 4894 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:55 crc kubenswrapper[4894]: I1209 15:35:55.135436 4894 status_manager.go:851] "Failed to get status for pod" podUID="5498bfbc-4945-4ffc-a7e3-db254f6cbabb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:55 crc kubenswrapper[4894]: I1209 15:35:55.135675 4894 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="5d56872e5289dd9a10f062411be41ee47e65db7b933678107f94ea3143c5238d" exitCode=0 Dec 09 15:35:55 crc kubenswrapper[4894]: I1209 15:35:55.135715 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"5d56872e5289dd9a10f062411be41ee47e65db7b933678107f94ea3143c5238d"} Dec 09 15:35:55 crc kubenswrapper[4894]: I1209 15:35:55.135725 4894 status_manager.go:851] "Failed to get status for pod" podUID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" pod="openshift-marketplace/redhat-operators-4bsxc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-4bsxc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:55 crc kubenswrapper[4894]: I1209 15:35:55.135991 4894 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9b8358d3-529c-43c9-964c-fe3bae20f1f8" Dec 09 15:35:55 crc kubenswrapper[4894]: I1209 15:35:55.136009 4894 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9b8358d3-529c-43c9-964c-fe3bae20f1f8" Dec 09 15:35:55 crc kubenswrapper[4894]: E1209 15:35:55.136402 4894 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:35:55 crc kubenswrapper[4894]: I1209 15:35:55.136603 4894 status_manager.go:851] "Failed to get status for pod" podUID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" pod="openshift-marketplace/redhat-marketplace-q58q9" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-marketplace-q58q9\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:55 crc kubenswrapper[4894]: I1209 15:35:55.136813 4894 status_manager.go:851] "Failed to get status for pod" podUID="f614b9022728cf315e60c057852e563e" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-controller-manager/pods/kube-controller-manager-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:55 crc kubenswrapper[4894]: I1209 15:35:55.136979 4894 status_manager.go:851] "Failed to get status for pod" podUID="5498bfbc-4945-4ffc-a7e3-db254f6cbabb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:55 crc kubenswrapper[4894]: I1209 15:35:55.137257 4894 status_manager.go:851] "Failed to get status for pod" podUID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" pod="openshift-marketplace/redhat-operators-4bsxc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/pods/redhat-operators-4bsxc\": dial tcp 38.102.83.64:6443: connect: connection refused" Dec 09 15:35:56 crc kubenswrapper[4894]: I1209 15:35:56.143109 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"c0ea4a5f90bc996c762c20014ec9264a4a8143f6a42e1f58ac77ed2ab2956f70"} Dec 09 15:35:56 crc kubenswrapper[4894]: I1209 15:35:56.144012 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"7158bdefdd0be3483fd1854b3675fde33732f293098abdab447931f998e079fd"} Dec 09 15:35:57 crc kubenswrapper[4894]: I1209 15:35:57.149935 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3804c37c62e2b0720dee0c20094cb55bad2284b8cbd1fe332f0527a441610e9c"} Dec 09 15:35:57 crc kubenswrapper[4894]: I1209 15:35:57.149982 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"69ec5814da6a33b76be9d15dbfba0ee42de45af1ec4d3f7820eef619c90d97fd"} Dec 09 15:35:57 crc kubenswrapper[4894]: I1209 15:35:57.149995 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2f2abf1d4d0c7f8064a18c91528b7569b18550a966326566e9fa1a5cac45d2a6"} Dec 09 15:35:57 crc kubenswrapper[4894]: I1209 15:35:57.150262 4894 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9b8358d3-529c-43c9-964c-fe3bae20f1f8" Dec 09 15:35:57 crc kubenswrapper[4894]: I1209 15:35:57.150278 4894 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9b8358d3-529c-43c9-964c-fe3bae20f1f8" Dec 09 15:35:57 crc kubenswrapper[4894]: I1209 15:35:57.150476 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:35:58 crc kubenswrapper[4894]: I1209 15:35:58.127284 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:35:58 crc kubenswrapper[4894]: I1209 15:35:58.127591 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:35:58 crc kubenswrapper[4894]: I1209 15:35:58.131615 4894 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]log ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]etcd ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/openshift.io-api-request-count-filter ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/openshift.io-startkubeinformers ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/start-apiserver-admission-initializer ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/generic-apiserver-start-informers ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/priority-and-fairness-config-consumer ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/priority-and-fairness-filter ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/storage-object-count-tracker-hook ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/start-apiextensions-informers ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/start-apiextensions-controllers ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/crd-informer-synced ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/start-system-namespaces-controller ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/start-cluster-authentication-info-controller ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/start-legacy-token-tracking-controller ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/start-service-ip-repair-controllers ok Dec 09 15:35:58 crc kubenswrapper[4894]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/priority-and-fairness-config-producer ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/bootstrap-controller ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/start-kube-aggregator-informers ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/apiservice-status-local-available-controller ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/apiservice-status-remote-available-controller ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/apiservice-registration-controller ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/apiservice-wait-for-first-sync ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/apiservice-discovery-controller ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/kube-apiserver-autoregistration ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]autoregister-completion ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/apiservice-openapi-controller ok Dec 09 15:35:58 crc kubenswrapper[4894]: [+]poststarthook/apiservice-openapiv3-controller ok Dec 09 15:35:58 crc kubenswrapper[4894]: livez check failed Dec 09 15:35:58 crc kubenswrapper[4894]: I1209 15:35:58.131669 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="71bb4a3aecc4ba5b26c4b7318770ce13" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Dec 09 15:35:58 crc kubenswrapper[4894]: I1209 15:35:58.356910 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 15:35:58 crc kubenswrapper[4894]: I1209 15:35:58.359995 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 15:35:59 crc kubenswrapper[4894]: I1209 15:35:59.159783 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 15:36:02 crc kubenswrapper[4894]: I1209 15:36:02.160420 4894 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:36:03 crc kubenswrapper[4894]: I1209 15:36:03.134498 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:36:03 crc kubenswrapper[4894]: I1209 15:36:03.138257 4894 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="b7536047-0777-42d0-84fb-0e5402baeb13" Dec 09 15:36:03 crc kubenswrapper[4894]: I1209 15:36:03.180220 4894 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9b8358d3-529c-43c9-964c-fe3bae20f1f8" Dec 09 15:36:03 crc kubenswrapper[4894]: I1209 15:36:03.180249 4894 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9b8358d3-529c-43c9-964c-fe3bae20f1f8" Dec 09 15:36:03 crc kubenswrapper[4894]: I1209 15:36:03.184718 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:36:04 crc kubenswrapper[4894]: I1209 15:36:04.185529 4894 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9b8358d3-529c-43c9-964c-fe3bae20f1f8" Dec 09 15:36:04 crc kubenswrapper[4894]: I1209 15:36:04.185576 4894 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="9b8358d3-529c-43c9-964c-fe3bae20f1f8" Dec 09 15:36:06 crc kubenswrapper[4894]: I1209 15:36:06.115547 4894 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="b7536047-0777-42d0-84fb-0e5402baeb13" Dec 09 15:36:09 crc kubenswrapper[4894]: I1209 15:36:09.763319 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Dec 09 15:36:12 crc kubenswrapper[4894]: I1209 15:36:12.071515 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Dec 09 15:36:12 crc kubenswrapper[4894]: I1209 15:36:12.560028 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Dec 09 15:36:12 crc kubenswrapper[4894]: I1209 15:36:12.713172 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Dec 09 15:36:12 crc kubenswrapper[4894]: I1209 15:36:12.731108 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Dec 09 15:36:12 crc kubenswrapper[4894]: I1209 15:36:12.825452 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Dec 09 15:36:13 crc kubenswrapper[4894]: I1209 15:36:13.088564 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Dec 09 15:36:13 crc kubenswrapper[4894]: I1209 15:36:13.124848 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Dec 09 15:36:13 crc kubenswrapper[4894]: I1209 15:36:13.187015 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Dec 09 15:36:13 crc kubenswrapper[4894]: I1209 15:36:13.470190 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Dec 09 15:36:13 crc kubenswrapper[4894]: I1209 15:36:13.667348 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Dec 09 15:36:13 crc kubenswrapper[4894]: I1209 15:36:13.800496 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Dec 09 15:36:13 crc kubenswrapper[4894]: I1209 15:36:13.800571 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Dec 09 15:36:13 crc kubenswrapper[4894]: I1209 15:36:13.874481 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Dec 09 15:36:14 crc kubenswrapper[4894]: I1209 15:36:14.138348 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Dec 09 15:36:14 crc kubenswrapper[4894]: I1209 15:36:14.239103 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Dec 09 15:36:14 crc kubenswrapper[4894]: I1209 15:36:14.316955 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Dec 09 15:36:14 crc kubenswrapper[4894]: I1209 15:36:14.343138 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Dec 09 15:36:14 crc kubenswrapper[4894]: I1209 15:36:14.455526 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Dec 09 15:36:14 crc kubenswrapper[4894]: I1209 15:36:14.485622 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Dec 09 15:36:14 crc kubenswrapper[4894]: I1209 15:36:14.607057 4894 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Dec 09 15:36:14 crc kubenswrapper[4894]: I1209 15:36:14.610371 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Dec 09 15:36:14 crc kubenswrapper[4894]: I1209 15:36:14.612212 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 09 15:36:14 crc kubenswrapper[4894]: I1209 15:36:14.612267 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Dec 09 15:36:14 crc kubenswrapper[4894]: I1209 15:36:14.616333 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Dec 09 15:36:14 crc kubenswrapper[4894]: I1209 15:36:14.628334 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=12.628313968 podStartE2EDuration="12.628313968s" podCreationTimestamp="2025-12-09 15:36:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:36:14.626385867 +0000 UTC m=+268.945596556" watchObservedRunningTime="2025-12-09 15:36:14.628313968 +0000 UTC m=+268.947524647" Dec 09 15:36:14 crc kubenswrapper[4894]: I1209 15:36:14.758306 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Dec 09 15:36:14 crc kubenswrapper[4894]: I1209 15:36:14.892344 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Dec 09 15:36:15 crc kubenswrapper[4894]: I1209 15:36:15.055664 4894 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Dec 09 15:36:15 crc kubenswrapper[4894]: I1209 15:36:15.105855 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Dec 09 15:36:15 crc kubenswrapper[4894]: I1209 15:36:15.172098 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Dec 09 15:36:15 crc kubenswrapper[4894]: I1209 15:36:15.201310 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Dec 09 15:36:15 crc kubenswrapper[4894]: I1209 15:36:15.250724 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Dec 09 15:36:15 crc kubenswrapper[4894]: I1209 15:36:15.463882 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Dec 09 15:36:15 crc kubenswrapper[4894]: I1209 15:36:15.490742 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Dec 09 15:36:15 crc kubenswrapper[4894]: I1209 15:36:15.505243 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Dec 09 15:36:15 crc kubenswrapper[4894]: I1209 15:36:15.568283 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Dec 09 15:36:15 crc kubenswrapper[4894]: I1209 15:36:15.657984 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Dec 09 15:36:15 crc kubenswrapper[4894]: I1209 15:36:15.716054 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Dec 09 15:36:15 crc kubenswrapper[4894]: I1209 15:36:15.769203 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Dec 09 15:36:15 crc kubenswrapper[4894]: I1209 15:36:15.794359 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 09 15:36:15 crc kubenswrapper[4894]: I1209 15:36:15.822887 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 09 15:36:15 crc kubenswrapper[4894]: I1209 15:36:15.863666 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Dec 09 15:36:15 crc kubenswrapper[4894]: I1209 15:36:15.889500 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Dec 09 15:36:15 crc kubenswrapper[4894]: I1209 15:36:15.962007 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Dec 09 15:36:16 crc kubenswrapper[4894]: I1209 15:36:16.039553 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Dec 09 15:36:16 crc kubenswrapper[4894]: I1209 15:36:16.050586 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Dec 09 15:36:16 crc kubenswrapper[4894]: I1209 15:36:16.075963 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Dec 09 15:36:16 crc kubenswrapper[4894]: I1209 15:36:16.090681 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Dec 09 15:36:16 crc kubenswrapper[4894]: I1209 15:36:16.199545 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 09 15:36:16 crc kubenswrapper[4894]: I1209 15:36:16.220500 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Dec 09 15:36:16 crc kubenswrapper[4894]: I1209 15:36:16.394677 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Dec 09 15:36:16 crc kubenswrapper[4894]: I1209 15:36:16.707277 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Dec 09 15:36:16 crc kubenswrapper[4894]: I1209 15:36:16.721359 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Dec 09 15:36:16 crc kubenswrapper[4894]: I1209 15:36:16.775945 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Dec 09 15:36:16 crc kubenswrapper[4894]: I1209 15:36:16.793347 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Dec 09 15:36:16 crc kubenswrapper[4894]: I1209 15:36:16.816631 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Dec 09 15:36:16 crc kubenswrapper[4894]: I1209 15:36:16.841774 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Dec 09 15:36:16 crc kubenswrapper[4894]: I1209 15:36:16.876021 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Dec 09 15:36:16 crc kubenswrapper[4894]: I1209 15:36:16.948371 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Dec 09 15:36:16 crc kubenswrapper[4894]: I1209 15:36:16.982065 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Dec 09 15:36:16 crc kubenswrapper[4894]: I1209 15:36:16.986736 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Dec 09 15:36:16 crc kubenswrapper[4894]: I1209 15:36:16.990791 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.006119 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.160346 4894 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.176329 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.193726 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.285123 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.351760 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.353906 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.364277 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.421881 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.483614 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.523530 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.606963 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.655532 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.695064 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.697503 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.744477 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.751479 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.764044 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.792862 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.805364 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.828876 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.899398 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.937191 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.958301 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.971268 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Dec 09 15:36:17 crc kubenswrapper[4894]: I1209 15:36:17.992968 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Dec 09 15:36:18 crc kubenswrapper[4894]: I1209 15:36:18.026079 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Dec 09 15:36:18 crc kubenswrapper[4894]: I1209 15:36:18.042032 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Dec 09 15:36:18 crc kubenswrapper[4894]: I1209 15:36:18.065702 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Dec 09 15:36:18 crc kubenswrapper[4894]: I1209 15:36:18.067565 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Dec 09 15:36:18 crc kubenswrapper[4894]: I1209 15:36:18.078590 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Dec 09 15:36:18 crc kubenswrapper[4894]: I1209 15:36:18.085456 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Dec 09 15:36:18 crc kubenswrapper[4894]: I1209 15:36:18.092169 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Dec 09 15:36:18 crc kubenswrapper[4894]: I1209 15:36:18.157571 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Dec 09 15:36:18 crc kubenswrapper[4894]: I1209 15:36:18.176596 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Dec 09 15:36:18 crc kubenswrapper[4894]: I1209 15:36:18.254720 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Dec 09 15:36:18 crc kubenswrapper[4894]: I1209 15:36:18.316547 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Dec 09 15:36:18 crc kubenswrapper[4894]: I1209 15:36:18.362109 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 09 15:36:18 crc kubenswrapper[4894]: I1209 15:36:18.505202 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Dec 09 15:36:18 crc kubenswrapper[4894]: I1209 15:36:18.535572 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Dec 09 15:36:18 crc kubenswrapper[4894]: I1209 15:36:18.631252 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Dec 09 15:36:18 crc kubenswrapper[4894]: I1209 15:36:18.710090 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Dec 09 15:36:18 crc kubenswrapper[4894]: I1209 15:36:18.751893 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Dec 09 15:36:18 crc kubenswrapper[4894]: I1209 15:36:18.770227 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Dec 09 15:36:19 crc kubenswrapper[4894]: I1209 15:36:19.068686 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Dec 09 15:36:19 crc kubenswrapper[4894]: I1209 15:36:19.114344 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Dec 09 15:36:19 crc kubenswrapper[4894]: I1209 15:36:19.145830 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Dec 09 15:36:19 crc kubenswrapper[4894]: I1209 15:36:19.184553 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Dec 09 15:36:19 crc kubenswrapper[4894]: I1209 15:36:19.396338 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Dec 09 15:36:19 crc kubenswrapper[4894]: I1209 15:36:19.458942 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Dec 09 15:36:19 crc kubenswrapper[4894]: I1209 15:36:19.461213 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Dec 09 15:36:19 crc kubenswrapper[4894]: I1209 15:36:19.462557 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Dec 09 15:36:19 crc kubenswrapper[4894]: I1209 15:36:19.479631 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Dec 09 15:36:19 crc kubenswrapper[4894]: I1209 15:36:19.563436 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Dec 09 15:36:19 crc kubenswrapper[4894]: I1209 15:36:19.625357 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Dec 09 15:36:19 crc kubenswrapper[4894]: I1209 15:36:19.626688 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Dec 09 15:36:19 crc kubenswrapper[4894]: I1209 15:36:19.656440 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 09 15:36:19 crc kubenswrapper[4894]: I1209 15:36:19.689927 4894 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Dec 09 15:36:19 crc kubenswrapper[4894]: I1209 15:36:19.697865 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Dec 09 15:36:19 crc kubenswrapper[4894]: I1209 15:36:19.718046 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Dec 09 15:36:19 crc kubenswrapper[4894]: I1209 15:36:19.819255 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Dec 09 15:36:19 crc kubenswrapper[4894]: I1209 15:36:19.821134 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Dec 09 15:36:19 crc kubenswrapper[4894]: I1209 15:36:19.907188 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Dec 09 15:36:19 crc kubenswrapper[4894]: I1209 15:36:19.938666 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 09 15:36:20 crc kubenswrapper[4894]: I1209 15:36:20.025675 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Dec 09 15:36:20 crc kubenswrapper[4894]: I1209 15:36:20.034335 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 09 15:36:20 crc kubenswrapper[4894]: I1209 15:36:20.192728 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Dec 09 15:36:20 crc kubenswrapper[4894]: I1209 15:36:20.239885 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Dec 09 15:36:20 crc kubenswrapper[4894]: I1209 15:36:20.262084 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Dec 09 15:36:20 crc kubenswrapper[4894]: I1209 15:36:20.456893 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 09 15:36:20 crc kubenswrapper[4894]: I1209 15:36:20.514691 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Dec 09 15:36:20 crc kubenswrapper[4894]: I1209 15:36:20.634998 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Dec 09 15:36:20 crc kubenswrapper[4894]: I1209 15:36:20.635742 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Dec 09 15:36:20 crc kubenswrapper[4894]: I1209 15:36:20.636020 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Dec 09 15:36:20 crc kubenswrapper[4894]: I1209 15:36:20.683549 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Dec 09 15:36:20 crc kubenswrapper[4894]: I1209 15:36:20.737680 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Dec 09 15:36:20 crc kubenswrapper[4894]: I1209 15:36:20.789730 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Dec 09 15:36:20 crc kubenswrapper[4894]: I1209 15:36:20.793078 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Dec 09 15:36:20 crc kubenswrapper[4894]: I1209 15:36:20.826367 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Dec 09 15:36:20 crc kubenswrapper[4894]: I1209 15:36:20.841346 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Dec 09 15:36:21 crc kubenswrapper[4894]: I1209 15:36:21.047037 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Dec 09 15:36:21 crc kubenswrapper[4894]: I1209 15:36:21.120001 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Dec 09 15:36:21 crc kubenswrapper[4894]: I1209 15:36:21.133390 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Dec 09 15:36:21 crc kubenswrapper[4894]: I1209 15:36:21.199150 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Dec 09 15:36:21 crc kubenswrapper[4894]: I1209 15:36:21.603204 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Dec 09 15:36:21 crc kubenswrapper[4894]: I1209 15:36:21.756807 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Dec 09 15:36:21 crc kubenswrapper[4894]: I1209 15:36:21.804191 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Dec 09 15:36:21 crc kubenswrapper[4894]: I1209 15:36:21.929889 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Dec 09 15:36:21 crc kubenswrapper[4894]: I1209 15:36:21.963408 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Dec 09 15:36:22 crc kubenswrapper[4894]: I1209 15:36:22.081796 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Dec 09 15:36:22 crc kubenswrapper[4894]: I1209 15:36:22.082407 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Dec 09 15:36:22 crc kubenswrapper[4894]: I1209 15:36:22.140759 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Dec 09 15:36:22 crc kubenswrapper[4894]: I1209 15:36:22.158110 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Dec 09 15:36:22 crc kubenswrapper[4894]: I1209 15:36:22.170769 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Dec 09 15:36:22 crc kubenswrapper[4894]: I1209 15:36:22.192537 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Dec 09 15:36:22 crc kubenswrapper[4894]: I1209 15:36:22.297347 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Dec 09 15:36:22 crc kubenswrapper[4894]: I1209 15:36:22.430477 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Dec 09 15:36:22 crc kubenswrapper[4894]: I1209 15:36:22.555423 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Dec 09 15:36:22 crc kubenswrapper[4894]: I1209 15:36:22.564015 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Dec 09 15:36:22 crc kubenswrapper[4894]: I1209 15:36:22.626306 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Dec 09 15:36:22 crc kubenswrapper[4894]: I1209 15:36:22.626624 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Dec 09 15:36:22 crc kubenswrapper[4894]: I1209 15:36:22.659939 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Dec 09 15:36:22 crc kubenswrapper[4894]: I1209 15:36:22.696136 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Dec 09 15:36:22 crc kubenswrapper[4894]: I1209 15:36:22.744259 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Dec 09 15:36:22 crc kubenswrapper[4894]: I1209 15:36:22.803069 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Dec 09 15:36:22 crc kubenswrapper[4894]: I1209 15:36:22.827219 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 09 15:36:22 crc kubenswrapper[4894]: I1209 15:36:22.829173 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Dec 09 15:36:22 crc kubenswrapper[4894]: I1209 15:36:22.917727 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Dec 09 15:36:22 crc kubenswrapper[4894]: I1209 15:36:22.946455 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Dec 09 15:36:22 crc kubenswrapper[4894]: I1209 15:36:22.959793 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.025982 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.092673 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.108302 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.124750 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.189818 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.317876 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.349563 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.440169 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.453123 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.484023 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.594783 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.612393 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.613397 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.619249 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.624446 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.646322 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.697323 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.710558 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.736592 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.743573 4894 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.780006 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.781024 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.797220 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.834457 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.953171 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Dec 09 15:36:23 crc kubenswrapper[4894]: I1209 15:36:23.956835 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Dec 09 15:36:24 crc kubenswrapper[4894]: I1209 15:36:24.002511 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Dec 09 15:36:24 crc kubenswrapper[4894]: I1209 15:36:24.152889 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Dec 09 15:36:24 crc kubenswrapper[4894]: I1209 15:36:24.167742 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Dec 09 15:36:24 crc kubenswrapper[4894]: I1209 15:36:24.190244 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Dec 09 15:36:24 crc kubenswrapper[4894]: I1209 15:36:24.190818 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Dec 09 15:36:24 crc kubenswrapper[4894]: I1209 15:36:24.225268 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Dec 09 15:36:24 crc kubenswrapper[4894]: I1209 15:36:24.249069 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Dec 09 15:36:24 crc kubenswrapper[4894]: I1209 15:36:24.310962 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Dec 09 15:36:24 crc kubenswrapper[4894]: I1209 15:36:24.400065 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Dec 09 15:36:24 crc kubenswrapper[4894]: I1209 15:36:24.430996 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Dec 09 15:36:24 crc kubenswrapper[4894]: I1209 15:36:24.457191 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Dec 09 15:36:24 crc kubenswrapper[4894]: I1209 15:36:24.530858 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Dec 09 15:36:24 crc kubenswrapper[4894]: I1209 15:36:24.547310 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Dec 09 15:36:24 crc kubenswrapper[4894]: I1209 15:36:24.626339 4894 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Dec 09 15:36:24 crc kubenswrapper[4894]: I1209 15:36:24.626680 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://051cca6ef214542ae81693b1c1177e599eba5687a444956d22beb2a398b4bbe3" gracePeriod=5 Dec 09 15:36:24 crc kubenswrapper[4894]: I1209 15:36:24.673039 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Dec 09 15:36:24 crc kubenswrapper[4894]: I1209 15:36:24.673133 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Dec 09 15:36:24 crc kubenswrapper[4894]: I1209 15:36:24.711784 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Dec 09 15:36:24 crc kubenswrapper[4894]: I1209 15:36:24.711929 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 09 15:36:24 crc kubenswrapper[4894]: I1209 15:36:24.765252 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 09 15:36:24 crc kubenswrapper[4894]: I1209 15:36:24.965975 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Dec 09 15:36:24 crc kubenswrapper[4894]: I1209 15:36:24.990433 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.068458 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.174267 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.213205 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hq8h6"] Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.214188 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hq8h6" podUID="4397f08d-278a-4e2d-9664-e44161c96638" containerName="registry-server" containerID="cri-o://f99ea2eb94afcf5102569043baa964d088bfa60c0648c833f9019a560c201378" gracePeriod=30 Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.227504 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ff286"] Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.227921 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-ff286" podUID="c278157d-d806-4fa1-a597-076d95a286e2" containerName="registry-server" containerID="cri-o://a15756c0576e9c5301de22221f66cec7e5bdb908775678f43f6efeeb8dac247e" gracePeriod=30 Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.232718 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qq9k2"] Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.233028 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" podUID="0847bb67-0cb7-4616-b715-29b6f12f67b5" containerName="marketplace-operator" containerID="cri-o://a255986c888b8d5cf49c452d7b4f3062861ed7cb28f3e965e92b1d5428262489" gracePeriod=30 Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.238874 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nxlqt"] Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.240126 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-nxlqt" podUID="2421584b-8a4b-4cb1-97c0-cb41d8ec16b2" containerName="registry-server" containerID="cri-o://0a2bf7f94932408830073b3ab94b03ba306028cef6f083fe1fa991aad52155fb" gracePeriod=30 Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.248297 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q58q9"] Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.248590 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q58q9" podUID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" containerName="registry-server" containerID="cri-o://6887e8e0a215db3554aceb2634a1e253366e560bf7d5cde02b24ee31510ac3bb" gracePeriod=30 Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.255381 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4bsxc"] Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.255598 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4bsxc" podUID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" containerName="registry-server" containerID="cri-o://760697f2531f933eeec09e43003f58c82b1077c976a4af23b6f5ae395cdd8a8a" gracePeriod=30 Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.285066 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v5dgx"] Dec 09 15:36:25 crc kubenswrapper[4894]: E1209 15:36:25.285660 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.285681 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 09 15:36:25 crc kubenswrapper[4894]: E1209 15:36:25.285702 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5498bfbc-4945-4ffc-a7e3-db254f6cbabb" containerName="installer" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.285711 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5498bfbc-4945-4ffc-a7e3-db254f6cbabb" containerName="installer" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.285863 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="5498bfbc-4945-4ffc-a7e3-db254f6cbabb" containerName="installer" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.285886 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.286384 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-v5dgx" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.296973 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v5dgx"] Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.323441 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.364993 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.373036 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.380486 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/02757ffa-44f3-4f5d-ad50-7d7044ce3a24-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-v5dgx\" (UID: \"02757ffa-44f3-4f5d-ad50-7d7044ce3a24\") " pod="openshift-marketplace/marketplace-operator-79b997595-v5dgx" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.380539 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/02757ffa-44f3-4f5d-ad50-7d7044ce3a24-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-v5dgx\" (UID: \"02757ffa-44f3-4f5d-ad50-7d7044ce3a24\") " pod="openshift-marketplace/marketplace-operator-79b997595-v5dgx" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.380576 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8pfsz\" (UniqueName: \"kubernetes.io/projected/02757ffa-44f3-4f5d-ad50-7d7044ce3a24-kube-api-access-8pfsz\") pod \"marketplace-operator-79b997595-v5dgx\" (UID: \"02757ffa-44f3-4f5d-ad50-7d7044ce3a24\") " pod="openshift-marketplace/marketplace-operator-79b997595-v5dgx" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.436726 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.481766 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8pfsz\" (UniqueName: \"kubernetes.io/projected/02757ffa-44f3-4f5d-ad50-7d7044ce3a24-kube-api-access-8pfsz\") pod \"marketplace-operator-79b997595-v5dgx\" (UID: \"02757ffa-44f3-4f5d-ad50-7d7044ce3a24\") " pod="openshift-marketplace/marketplace-operator-79b997595-v5dgx" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.481876 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/02757ffa-44f3-4f5d-ad50-7d7044ce3a24-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-v5dgx\" (UID: \"02757ffa-44f3-4f5d-ad50-7d7044ce3a24\") " pod="openshift-marketplace/marketplace-operator-79b997595-v5dgx" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.481910 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/02757ffa-44f3-4f5d-ad50-7d7044ce3a24-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-v5dgx\" (UID: \"02757ffa-44f3-4f5d-ad50-7d7044ce3a24\") " pod="openshift-marketplace/marketplace-operator-79b997595-v5dgx" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.483391 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/02757ffa-44f3-4f5d-ad50-7d7044ce3a24-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-v5dgx\" (UID: \"02757ffa-44f3-4f5d-ad50-7d7044ce3a24\") " pod="openshift-marketplace/marketplace-operator-79b997595-v5dgx" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.490979 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/02757ffa-44f3-4f5d-ad50-7d7044ce3a24-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-v5dgx\" (UID: \"02757ffa-44f3-4f5d-ad50-7d7044ce3a24\") " pod="openshift-marketplace/marketplace-operator-79b997595-v5dgx" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.509483 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8pfsz\" (UniqueName: \"kubernetes.io/projected/02757ffa-44f3-4f5d-ad50-7d7044ce3a24-kube-api-access-8pfsz\") pod \"marketplace-operator-79b997595-v5dgx\" (UID: \"02757ffa-44f3-4f5d-ad50-7d7044ce3a24\") " pod="openshift-marketplace/marketplace-operator-79b997595-v5dgx" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.560166 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.603946 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-v5dgx" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.682486 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.750021 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.801855 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-v5dgx"] Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.883472 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.921095 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Dec 09 15:36:25 crc kubenswrapper[4894]: I1209 15:36:25.945261 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Dec 09 15:36:26 crc kubenswrapper[4894]: I1209 15:36:26.052531 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Dec 09 15:36:26 crc kubenswrapper[4894]: I1209 15:36:26.111678 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Dec 09 15:36:26 crc kubenswrapper[4894]: I1209 15:36:26.146680 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Dec 09 15:36:26 crc kubenswrapper[4894]: I1209 15:36:26.294182 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-v5dgx" event={"ID":"02757ffa-44f3-4f5d-ad50-7d7044ce3a24","Type":"ContainerStarted","Data":"379b7d7d31e7c5b9b1127301af4d98cadd98eb0f224687de3577b0fe98a46ae0"} Dec 09 15:36:26 crc kubenswrapper[4894]: I1209 15:36:26.428578 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Dec 09 15:36:26 crc kubenswrapper[4894]: I1209 15:36:26.550183 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Dec 09 15:36:26 crc kubenswrapper[4894]: I1209 15:36:26.655763 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Dec 09 15:36:26 crc kubenswrapper[4894]: I1209 15:36:26.738375 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Dec 09 15:36:26 crc kubenswrapper[4894]: I1209 15:36:26.740499 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Dec 09 15:36:26 crc kubenswrapper[4894]: I1209 15:36:26.742858 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 09 15:36:26 crc kubenswrapper[4894]: I1209 15:36:26.763965 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Dec 09 15:36:26 crc kubenswrapper[4894]: I1209 15:36:26.849608 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Dec 09 15:36:26 crc kubenswrapper[4894]: I1209 15:36:26.971201 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Dec 09 15:36:26 crc kubenswrapper[4894]: I1209 15:36:26.991474 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ff286" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.090186 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.104480 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4d2d\" (UniqueName: \"kubernetes.io/projected/c278157d-d806-4fa1-a597-076d95a286e2-kube-api-access-r4d2d\") pod \"c278157d-d806-4fa1-a597-076d95a286e2\" (UID: \"c278157d-d806-4fa1-a597-076d95a286e2\") " Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.104615 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c278157d-d806-4fa1-a597-076d95a286e2-catalog-content\") pod \"c278157d-d806-4fa1-a597-076d95a286e2\" (UID: \"c278157d-d806-4fa1-a597-076d95a286e2\") " Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.104673 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c278157d-d806-4fa1-a597-076d95a286e2-utilities\") pod \"c278157d-d806-4fa1-a597-076d95a286e2\" (UID: \"c278157d-d806-4fa1-a597-076d95a286e2\") " Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.105834 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c278157d-d806-4fa1-a597-076d95a286e2-utilities" (OuterVolumeSpecName: "utilities") pod "c278157d-d806-4fa1-a597-076d95a286e2" (UID: "c278157d-d806-4fa1-a597-076d95a286e2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.116170 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.121042 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c278157d-d806-4fa1-a597-076d95a286e2-kube-api-access-r4d2d" (OuterVolumeSpecName: "kube-api-access-r4d2d") pod "c278157d-d806-4fa1-a597-076d95a286e2" (UID: "c278157d-d806-4fa1-a597-076d95a286e2"). InnerVolumeSpecName "kube-api-access-r4d2d". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.132402 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.152334 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c278157d-d806-4fa1-a597-076d95a286e2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c278157d-d806-4fa1-a597-076d95a286e2" (UID: "c278157d-d806-4fa1-a597-076d95a286e2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.205765 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4d2d\" (UniqueName: \"kubernetes.io/projected/c278157d-d806-4fa1-a597-076d95a286e2-kube-api-access-r4d2d\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.205799 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c278157d-d806-4fa1-a597-076d95a286e2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.205808 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c278157d-d806-4fa1-a597-076d95a286e2-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.300604 4894 generic.go:334] "Generic (PLEG): container finished" podID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" containerID="6887e8e0a215db3554aceb2634a1e253366e560bf7d5cde02b24ee31510ac3bb" exitCode=0 Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.300686 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q58q9" event={"ID":"ee374eee-998a-48f2-acdf-39daa5ce2fdd","Type":"ContainerDied","Data":"6887e8e0a215db3554aceb2634a1e253366e560bf7d5cde02b24ee31510ac3bb"} Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.302256 4894 generic.go:334] "Generic (PLEG): container finished" podID="0847bb67-0cb7-4616-b715-29b6f12f67b5" containerID="a255986c888b8d5cf49c452d7b4f3062861ed7cb28f3e965e92b1d5428262489" exitCode=0 Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.302324 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" event={"ID":"0847bb67-0cb7-4616-b715-29b6f12f67b5","Type":"ContainerDied","Data":"a255986c888b8d5cf49c452d7b4f3062861ed7cb28f3e965e92b1d5428262489"} Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.304902 4894 generic.go:334] "Generic (PLEG): container finished" podID="c278157d-d806-4fa1-a597-076d95a286e2" containerID="a15756c0576e9c5301de22221f66cec7e5bdb908775678f43f6efeeb8dac247e" exitCode=0 Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.304972 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ff286" event={"ID":"c278157d-d806-4fa1-a597-076d95a286e2","Type":"ContainerDied","Data":"a15756c0576e9c5301de22221f66cec7e5bdb908775678f43f6efeeb8dac247e"} Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.304982 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-ff286" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.305006 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-ff286" event={"ID":"c278157d-d806-4fa1-a597-076d95a286e2","Type":"ContainerDied","Data":"8240059115ee9c7c5aa57f051668183b7740875bf4567a338cc19b74cce4febc"} Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.305032 4894 scope.go:117] "RemoveContainer" containerID="a15756c0576e9c5301de22221f66cec7e5bdb908775678f43f6efeeb8dac247e" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.306397 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-v5dgx" event={"ID":"02757ffa-44f3-4f5d-ad50-7d7044ce3a24","Type":"ContainerStarted","Data":"9db5d2d1fbc7e6a97e81fca1978b2be16e1b0ebd9542c65d5b290b7afb00a9cc"} Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.307668 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-v5dgx" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.311237 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-v5dgx" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.313087 4894 generic.go:334] "Generic (PLEG): container finished" podID="4397f08d-278a-4e2d-9664-e44161c96638" containerID="f99ea2eb94afcf5102569043baa964d088bfa60c0648c833f9019a560c201378" exitCode=0 Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.313154 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hq8h6" event={"ID":"4397f08d-278a-4e2d-9664-e44161c96638","Type":"ContainerDied","Data":"f99ea2eb94afcf5102569043baa964d088bfa60c0648c833f9019a560c201378"} Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.314901 4894 generic.go:334] "Generic (PLEG): container finished" podID="2421584b-8a4b-4cb1-97c0-cb41d8ec16b2" containerID="0a2bf7f94932408830073b3ab94b03ba306028cef6f083fe1fa991aad52155fb" exitCode=0 Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.314977 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxlqt" event={"ID":"2421584b-8a4b-4cb1-97c0-cb41d8ec16b2","Type":"ContainerDied","Data":"0a2bf7f94932408830073b3ab94b03ba306028cef6f083fe1fa991aad52155fb"} Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.316838 4894 generic.go:334] "Generic (PLEG): container finished" podID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" containerID="760697f2531f933eeec09e43003f58c82b1077c976a4af23b6f5ae395cdd8a8a" exitCode=0 Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.316863 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bsxc" event={"ID":"bfb801bd-e785-4fdc-8a34-d2b8769d3a29","Type":"ContainerDied","Data":"760697f2531f933eeec09e43003f58c82b1077c976a4af23b6f5ae395cdd8a8a"} Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.319725 4894 scope.go:117] "RemoveContainer" containerID="3dbbd9400817dcec45a38afefbd879bcf664787acea95df58a4d7763fd2f2d5a" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.326066 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-v5dgx" podStartSLOduration=2.326048432 podStartE2EDuration="2.326048432s" podCreationTimestamp="2025-12-09 15:36:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:36:27.320989185 +0000 UTC m=+281.640199854" watchObservedRunningTime="2025-12-09 15:36:27.326048432 +0000 UTC m=+281.645259101" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.374511 4894 scope.go:117] "RemoveContainer" containerID="7579df81a9780e5a1398c81e9f7ae75cb015cb52640413bdfb24749a8ea432a0" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.375581 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.377174 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-ff286"] Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.381785 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-ff286"] Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.412633 4894 scope.go:117] "RemoveContainer" containerID="a15756c0576e9c5301de22221f66cec7e5bdb908775678f43f6efeeb8dac247e" Dec 09 15:36:27 crc kubenswrapper[4894]: E1209 15:36:27.416785 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a15756c0576e9c5301de22221f66cec7e5bdb908775678f43f6efeeb8dac247e\": container with ID starting with a15756c0576e9c5301de22221f66cec7e5bdb908775678f43f6efeeb8dac247e not found: ID does not exist" containerID="a15756c0576e9c5301de22221f66cec7e5bdb908775678f43f6efeeb8dac247e" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.416830 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a15756c0576e9c5301de22221f66cec7e5bdb908775678f43f6efeeb8dac247e"} err="failed to get container status \"a15756c0576e9c5301de22221f66cec7e5bdb908775678f43f6efeeb8dac247e\": rpc error: code = NotFound desc = could not find container \"a15756c0576e9c5301de22221f66cec7e5bdb908775678f43f6efeeb8dac247e\": container with ID starting with a15756c0576e9c5301de22221f66cec7e5bdb908775678f43f6efeeb8dac247e not found: ID does not exist" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.416854 4894 scope.go:117] "RemoveContainer" containerID="3dbbd9400817dcec45a38afefbd879bcf664787acea95df58a4d7763fd2f2d5a" Dec 09 15:36:27 crc kubenswrapper[4894]: E1209 15:36:27.417848 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dbbd9400817dcec45a38afefbd879bcf664787acea95df58a4d7763fd2f2d5a\": container with ID starting with 3dbbd9400817dcec45a38afefbd879bcf664787acea95df58a4d7763fd2f2d5a not found: ID does not exist" containerID="3dbbd9400817dcec45a38afefbd879bcf664787acea95df58a4d7763fd2f2d5a" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.417876 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dbbd9400817dcec45a38afefbd879bcf664787acea95df58a4d7763fd2f2d5a"} err="failed to get container status \"3dbbd9400817dcec45a38afefbd879bcf664787acea95df58a4d7763fd2f2d5a\": rpc error: code = NotFound desc = could not find container \"3dbbd9400817dcec45a38afefbd879bcf664787acea95df58a4d7763fd2f2d5a\": container with ID starting with 3dbbd9400817dcec45a38afefbd879bcf664787acea95df58a4d7763fd2f2d5a not found: ID does not exist" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.417896 4894 scope.go:117] "RemoveContainer" containerID="7579df81a9780e5a1398c81e9f7ae75cb015cb52640413bdfb24749a8ea432a0" Dec 09 15:36:27 crc kubenswrapper[4894]: E1209 15:36:27.425784 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7579df81a9780e5a1398c81e9f7ae75cb015cb52640413bdfb24749a8ea432a0\": container with ID starting with 7579df81a9780e5a1398c81e9f7ae75cb015cb52640413bdfb24749a8ea432a0 not found: ID does not exist" containerID="7579df81a9780e5a1398c81e9f7ae75cb015cb52640413bdfb24749a8ea432a0" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.425834 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7579df81a9780e5a1398c81e9f7ae75cb015cb52640413bdfb24749a8ea432a0"} err="failed to get container status \"7579df81a9780e5a1398c81e9f7ae75cb015cb52640413bdfb24749a8ea432a0\": rpc error: code = NotFound desc = could not find container \"7579df81a9780e5a1398c81e9f7ae75cb015cb52640413bdfb24749a8ea432a0\": container with ID starting with 7579df81a9780e5a1398c81e9f7ae75cb015cb52640413bdfb24749a8ea432a0 not found: ID does not exist" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.515693 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.521301 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 09 15:36:27 crc kubenswrapper[4894]: I1209 15:36:27.642182 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:27.774145 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:27.961987 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:28.070622 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:28.111885 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c278157d-d806-4fa1-a597-076d95a286e2" path="/var/lib/kubelet/pods/c278157d-d806-4fa1-a597-076d95a286e2/volumes" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:28.245362 4894 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.490079 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.512048 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hq8h6" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.515745 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.520433 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nxlqt" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.525190 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q58q9" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.547021 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4397f08d-278a-4e2d-9664-e44161c96638-utilities\") pod \"4397f08d-278a-4e2d-9664-e44161c96638\" (UID: \"4397f08d-278a-4e2d-9664-e44161c96638\") " Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.547073 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0847bb67-0cb7-4616-b715-29b6f12f67b5-marketplace-trusted-ca\") pod \"0847bb67-0cb7-4616-b715-29b6f12f67b5\" (UID: \"0847bb67-0cb7-4616-b715-29b6f12f67b5\") " Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.547117 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4397f08d-278a-4e2d-9664-e44161c96638-catalog-content\") pod \"4397f08d-278a-4e2d-9664-e44161c96638\" (UID: \"4397f08d-278a-4e2d-9664-e44161c96638\") " Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.547140 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4k9sg\" (UniqueName: \"kubernetes.io/projected/4397f08d-278a-4e2d-9664-e44161c96638-kube-api-access-4k9sg\") pod \"4397f08d-278a-4e2d-9664-e44161c96638\" (UID: \"4397f08d-278a-4e2d-9664-e44161c96638\") " Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.547167 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2421584b-8a4b-4cb1-97c0-cb41d8ec16b2-utilities\") pod \"2421584b-8a4b-4cb1-97c0-cb41d8ec16b2\" (UID: \"2421584b-8a4b-4cb1-97c0-cb41d8ec16b2\") " Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.547192 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee374eee-998a-48f2-acdf-39daa5ce2fdd-catalog-content\") pod \"ee374eee-998a-48f2-acdf-39daa5ce2fdd\" (UID: \"ee374eee-998a-48f2-acdf-39daa5ce2fdd\") " Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.547220 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cd7hx\" (UniqueName: \"kubernetes.io/projected/0847bb67-0cb7-4616-b715-29b6f12f67b5-kube-api-access-cd7hx\") pod \"0847bb67-0cb7-4616-b715-29b6f12f67b5\" (UID: \"0847bb67-0cb7-4616-b715-29b6f12f67b5\") " Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.547238 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0847bb67-0cb7-4616-b715-29b6f12f67b5-marketplace-operator-metrics\") pod \"0847bb67-0cb7-4616-b715-29b6f12f67b5\" (UID: \"0847bb67-0cb7-4616-b715-29b6f12f67b5\") " Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.547283 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee374eee-998a-48f2-acdf-39daa5ce2fdd-utilities\") pod \"ee374eee-998a-48f2-acdf-39daa5ce2fdd\" (UID: \"ee374eee-998a-48f2-acdf-39daa5ce2fdd\") " Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.547302 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2421584b-8a4b-4cb1-97c0-cb41d8ec16b2-catalog-content\") pod \"2421584b-8a4b-4cb1-97c0-cb41d8ec16b2\" (UID: \"2421584b-8a4b-4cb1-97c0-cb41d8ec16b2\") " Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.551184 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee374eee-998a-48f2-acdf-39daa5ce2fdd-utilities" (OuterVolumeSpecName: "utilities") pod "ee374eee-998a-48f2-acdf-39daa5ce2fdd" (UID: "ee374eee-998a-48f2-acdf-39daa5ce2fdd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.551193 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-whpnq\" (UniqueName: \"kubernetes.io/projected/2421584b-8a4b-4cb1-97c0-cb41d8ec16b2-kube-api-access-whpnq\") pod \"2421584b-8a4b-4cb1-97c0-cb41d8ec16b2\" (UID: \"2421584b-8a4b-4cb1-97c0-cb41d8ec16b2\") " Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.551269 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hwmhn\" (UniqueName: \"kubernetes.io/projected/ee374eee-998a-48f2-acdf-39daa5ce2fdd-kube-api-access-hwmhn\") pod \"ee374eee-998a-48f2-acdf-39daa5ce2fdd\" (UID: \"ee374eee-998a-48f2-acdf-39daa5ce2fdd\") " Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.551584 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ee374eee-998a-48f2-acdf-39daa5ce2fdd-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.552625 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4397f08d-278a-4e2d-9664-e44161c96638-utilities" (OuterVolumeSpecName: "utilities") pod "4397f08d-278a-4e2d-9664-e44161c96638" (UID: "4397f08d-278a-4e2d-9664-e44161c96638"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.554035 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bsxc" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.557031 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0847bb67-0cb7-4616-b715-29b6f12f67b5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "0847bb67-0cb7-4616-b715-29b6f12f67b5" (UID: "0847bb67-0cb7-4616-b715-29b6f12f67b5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.557485 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2421584b-8a4b-4cb1-97c0-cb41d8ec16b2-utilities" (OuterVolumeSpecName: "utilities") pod "2421584b-8a4b-4cb1-97c0-cb41d8ec16b2" (UID: "2421584b-8a4b-4cb1-97c0-cb41d8ec16b2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.560621 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4397f08d-278a-4e2d-9664-e44161c96638-kube-api-access-4k9sg" (OuterVolumeSpecName: "kube-api-access-4k9sg") pod "4397f08d-278a-4e2d-9664-e44161c96638" (UID: "4397f08d-278a-4e2d-9664-e44161c96638"). InnerVolumeSpecName "kube-api-access-4k9sg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.567840 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee374eee-998a-48f2-acdf-39daa5ce2fdd-kube-api-access-hwmhn" (OuterVolumeSpecName: "kube-api-access-hwmhn") pod "ee374eee-998a-48f2-acdf-39daa5ce2fdd" (UID: "ee374eee-998a-48f2-acdf-39daa5ce2fdd"). InnerVolumeSpecName "kube-api-access-hwmhn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.569854 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0847bb67-0cb7-4616-b715-29b6f12f67b5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "0847bb67-0cb7-4616-b715-29b6f12f67b5" (UID: "0847bb67-0cb7-4616-b715-29b6f12f67b5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.571019 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0847bb67-0cb7-4616-b715-29b6f12f67b5-kube-api-access-cd7hx" (OuterVolumeSpecName: "kube-api-access-cd7hx") pod "0847bb67-0cb7-4616-b715-29b6f12f67b5" (UID: "0847bb67-0cb7-4616-b715-29b6f12f67b5"). InnerVolumeSpecName "kube-api-access-cd7hx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.578849 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2421584b-8a4b-4cb1-97c0-cb41d8ec16b2-kube-api-access-whpnq" (OuterVolumeSpecName: "kube-api-access-whpnq") pod "2421584b-8a4b-4cb1-97c0-cb41d8ec16b2" (UID: "2421584b-8a4b-4cb1-97c0-cb41d8ec16b2"). InnerVolumeSpecName "kube-api-access-whpnq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.586277 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2421584b-8a4b-4cb1-97c0-cb41d8ec16b2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2421584b-8a4b-4cb1-97c0-cb41d8ec16b2" (UID: "2421584b-8a4b-4cb1-97c0-cb41d8ec16b2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.591156 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ee374eee-998a-48f2-acdf-39daa5ce2fdd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ee374eee-998a-48f2-acdf-39daa5ce2fdd" (UID: "ee374eee-998a-48f2-acdf-39daa5ce2fdd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.608737 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4397f08d-278a-4e2d-9664-e44161c96638-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4397f08d-278a-4e2d-9664-e44161c96638" (UID: "4397f08d-278a-4e2d-9664-e44161c96638"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.652213 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfb801bd-e785-4fdc-8a34-d2b8769d3a29-utilities\") pod \"bfb801bd-e785-4fdc-8a34-d2b8769d3a29\" (UID: \"bfb801bd-e785-4fdc-8a34-d2b8769d3a29\") " Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.652586 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfb801bd-e785-4fdc-8a34-d2b8769d3a29-catalog-content\") pod \"bfb801bd-e785-4fdc-8a34-d2b8769d3a29\" (UID: \"bfb801bd-e785-4fdc-8a34-d2b8769d3a29\") " Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.653041 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfb801bd-e785-4fdc-8a34-d2b8769d3a29-utilities" (OuterVolumeSpecName: "utilities") pod "bfb801bd-e785-4fdc-8a34-d2b8769d3a29" (UID: "bfb801bd-e785-4fdc-8a34-d2b8769d3a29"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.653746 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-67ndf\" (UniqueName: \"kubernetes.io/projected/bfb801bd-e785-4fdc-8a34-d2b8769d3a29-kube-api-access-67ndf\") pod \"bfb801bd-e785-4fdc-8a34-d2b8769d3a29\" (UID: \"bfb801bd-e785-4fdc-8a34-d2b8769d3a29\") " Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.654023 4894 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/0847bb67-0cb7-4616-b715-29b6f12f67b5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.654060 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4397f08d-278a-4e2d-9664-e44161c96638-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.654074 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4k9sg\" (UniqueName: \"kubernetes.io/projected/4397f08d-278a-4e2d-9664-e44161c96638-kube-api-access-4k9sg\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.654087 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2421584b-8a4b-4cb1-97c0-cb41d8ec16b2-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.654098 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ee374eee-998a-48f2-acdf-39daa5ce2fdd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.654110 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cd7hx\" (UniqueName: \"kubernetes.io/projected/0847bb67-0cb7-4616-b715-29b6f12f67b5-kube-api-access-cd7hx\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.654123 4894 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/0847bb67-0cb7-4616-b715-29b6f12f67b5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.654135 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfb801bd-e785-4fdc-8a34-d2b8769d3a29-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.654146 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2421584b-8a4b-4cb1-97c0-cb41d8ec16b2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.654158 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-whpnq\" (UniqueName: \"kubernetes.io/projected/2421584b-8a4b-4cb1-97c0-cb41d8ec16b2-kube-api-access-whpnq\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.654171 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hwmhn\" (UniqueName: \"kubernetes.io/projected/ee374eee-998a-48f2-acdf-39daa5ce2fdd-kube-api-access-hwmhn\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.654182 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4397f08d-278a-4e2d-9664-e44161c96638-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.657780 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfb801bd-e785-4fdc-8a34-d2b8769d3a29-kube-api-access-67ndf" (OuterVolumeSpecName: "kube-api-access-67ndf") pod "bfb801bd-e785-4fdc-8a34-d2b8769d3a29" (UID: "bfb801bd-e785-4fdc-8a34-d2b8769d3a29"). InnerVolumeSpecName "kube-api-access-67ndf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.751274 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfb801bd-e785-4fdc-8a34-d2b8769d3a29-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bfb801bd-e785-4fdc-8a34-d2b8769d3a29" (UID: "bfb801bd-e785-4fdc-8a34-d2b8769d3a29"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.755446 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-67ndf\" (UniqueName: \"kubernetes.io/projected/bfb801bd-e785-4fdc-8a34-d2b8769d3a29-kube-api-access-67ndf\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:29 crc kubenswrapper[4894]: I1209 15:36:29.755468 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfb801bd-e785-4fdc-8a34-d2b8769d3a29-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.315223 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.315347 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.342894 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4bsxc" event={"ID":"bfb801bd-e785-4fdc-8a34-d2b8769d3a29","Type":"ContainerDied","Data":"425abcd653efcfd1107b3fd4179cf301058483398ea6aebdccf719c14c17a786"} Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.342978 4894 scope.go:117] "RemoveContainer" containerID="760697f2531f933eeec09e43003f58c82b1077c976a4af23b6f5ae395cdd8a8a" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.342921 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4bsxc" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.346870 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q58q9" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.347010 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q58q9" event={"ID":"ee374eee-998a-48f2-acdf-39daa5ce2fdd","Type":"ContainerDied","Data":"b2bf84540592cbaa8843f5b8d3839eebf0d3a1fb44ef2be442cffd16639cd089"} Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.349010 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.349078 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-qq9k2" event={"ID":"0847bb67-0cb7-4616-b715-29b6f12f67b5","Type":"ContainerDied","Data":"c3cf02ed9dedfecf73ef20da80c3e0c96e5f3e9f1f14f203c7041cc1cc97736d"} Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.350878 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.350966 4894 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="051cca6ef214542ae81693b1c1177e599eba5687a444956d22beb2a398b4bbe3" exitCode=137 Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.351067 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.357165 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hq8h6" event={"ID":"4397f08d-278a-4e2d-9664-e44161c96638","Type":"ContainerDied","Data":"00e10a54574bc8485c5bef256c44e50c8eb056dafe9840669ae7df957ec49776"} Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.357185 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hq8h6" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.360740 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.360787 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.360874 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.360899 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.361016 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.361046 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.361401 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-nxlqt" event={"ID":"2421584b-8a4b-4cb1-97c0-cb41d8ec16b2","Type":"ContainerDied","Data":"712254f1d9317e269ef0c61a6c33e1f85a6bb1273c6989b297bd36a8e9691081"} Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.361418 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-nxlqt" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.361475 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.361528 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.361650 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.362587 4894 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.362606 4894 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.362615 4894 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.362625 4894 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.369558 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.369809 4894 scope.go:117] "RemoveContainer" containerID="d36d2a68f74ea0b8f46707474565374cac5914f5a72ff6b285884a6db33d90ce" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.380737 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4bsxc"] Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.388761 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4bsxc"] Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.390296 4894 scope.go:117] "RemoveContainer" containerID="145672364d83611684af8e281cb8d9a14305f05659503f09298059c0eb54a7d1" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.393304 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q58q9"] Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.397161 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q58q9"] Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.401495 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qq9k2"] Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.404808 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-qq9k2"] Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.411835 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-nxlqt"] Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.416721 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-nxlqt"] Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.420013 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hq8h6"] Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.422593 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hq8h6"] Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.423224 4894 scope.go:117] "RemoveContainer" containerID="6887e8e0a215db3554aceb2634a1e253366e560bf7d5cde02b24ee31510ac3bb" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.454077 4894 scope.go:117] "RemoveContainer" containerID="c695feb4cb7d0efbde1fd625bdfb372189c361bd0398971d0af9488f048f090a" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.463657 4894 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.467778 4894 scope.go:117] "RemoveContainer" containerID="9ade3816e5bf9bba5a076fc5ba04468b688cdbcb09e029720868c42858aecfff" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.489854 4894 scope.go:117] "RemoveContainer" containerID="a255986c888b8d5cf49c452d7b4f3062861ed7cb28f3e965e92b1d5428262489" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.516935 4894 scope.go:117] "RemoveContainer" containerID="051cca6ef214542ae81693b1c1177e599eba5687a444956d22beb2a398b4bbe3" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.533728 4894 scope.go:117] "RemoveContainer" containerID="051cca6ef214542ae81693b1c1177e599eba5687a444956d22beb2a398b4bbe3" Dec 09 15:36:30 crc kubenswrapper[4894]: E1209 15:36:30.534248 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"051cca6ef214542ae81693b1c1177e599eba5687a444956d22beb2a398b4bbe3\": container with ID starting with 051cca6ef214542ae81693b1c1177e599eba5687a444956d22beb2a398b4bbe3 not found: ID does not exist" containerID="051cca6ef214542ae81693b1c1177e599eba5687a444956d22beb2a398b4bbe3" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.534293 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"051cca6ef214542ae81693b1c1177e599eba5687a444956d22beb2a398b4bbe3"} err="failed to get container status \"051cca6ef214542ae81693b1c1177e599eba5687a444956d22beb2a398b4bbe3\": rpc error: code = NotFound desc = could not find container \"051cca6ef214542ae81693b1c1177e599eba5687a444956d22beb2a398b4bbe3\": container with ID starting with 051cca6ef214542ae81693b1c1177e599eba5687a444956d22beb2a398b4bbe3 not found: ID does not exist" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.534325 4894 scope.go:117] "RemoveContainer" containerID="f99ea2eb94afcf5102569043baa964d088bfa60c0648c833f9019a560c201378" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.549869 4894 scope.go:117] "RemoveContainer" containerID="6226e3de7692ce8337bd03063725bc9bc03bae168a181c1cfb5e6022767f3cc1" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.570716 4894 scope.go:117] "RemoveContainer" containerID="cfba892241222e18b3fdce72ef9bb2cefd0c77fbcb02ba2d7ad49f5d06c56b48" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.592030 4894 scope.go:117] "RemoveContainer" containerID="0a2bf7f94932408830073b3ab94b03ba306028cef6f083fe1fa991aad52155fb" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.609104 4894 scope.go:117] "RemoveContainer" containerID="ca8726e6cc5f2a3db199c57f3943b724e01f4dd5bfcbe321417b335da1888b7a" Dec 09 15:36:30 crc kubenswrapper[4894]: I1209 15:36:30.621738 4894 scope.go:117] "RemoveContainer" containerID="6356f5ee31ced0642b88fc9d15acdb1b7ce35ffee04c3e156a38a54f45ac947c" Dec 09 15:36:32 crc kubenswrapper[4894]: I1209 15:36:32.113398 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0847bb67-0cb7-4616-b715-29b6f12f67b5" path="/var/lib/kubelet/pods/0847bb67-0cb7-4616-b715-29b6f12f67b5/volumes" Dec 09 15:36:32 crc kubenswrapper[4894]: I1209 15:36:32.114142 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2421584b-8a4b-4cb1-97c0-cb41d8ec16b2" path="/var/lib/kubelet/pods/2421584b-8a4b-4cb1-97c0-cb41d8ec16b2/volumes" Dec 09 15:36:32 crc kubenswrapper[4894]: I1209 15:36:32.114693 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4397f08d-278a-4e2d-9664-e44161c96638" path="/var/lib/kubelet/pods/4397f08d-278a-4e2d-9664-e44161c96638/volumes" Dec 09 15:36:32 crc kubenswrapper[4894]: I1209 15:36:32.115908 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" path="/var/lib/kubelet/pods/bfb801bd-e785-4fdc-8a34-d2b8769d3a29/volumes" Dec 09 15:36:32 crc kubenswrapper[4894]: I1209 15:36:32.116653 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" path="/var/lib/kubelet/pods/ee374eee-998a-48f2-acdf-39daa5ce2fdd/volumes" Dec 09 15:36:32 crc kubenswrapper[4894]: I1209 15:36:32.117553 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Dec 09 15:36:39 crc kubenswrapper[4894]: I1209 15:36:39.526513 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.012372 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qjlt6"] Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.013060 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" podUID="00d36531-cc42-49a2-a4c9-28f9a5ddd044" containerName="controller-manager" containerID="cri-o://8b7241b73e8a1ff5b13e5732d0d1c9ab20a45823a3e5454d8a97537de5a9680c" gracePeriod=30 Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.119501 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826"] Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.120223 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" podUID="171c863a-36f5-48c0-be81-752684537f0b" containerName="route-controller-manager" containerID="cri-o://79966100b093d6f879432f696bf6008f04634486cdb672b037d179f09313946f" gracePeriod=30 Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.370317 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.445237 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.453625 4894 generic.go:334] "Generic (PLEG): container finished" podID="00d36531-cc42-49a2-a4c9-28f9a5ddd044" containerID="8b7241b73e8a1ff5b13e5732d0d1c9ab20a45823a3e5454d8a97537de5a9680c" exitCode=0 Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.453715 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" event={"ID":"00d36531-cc42-49a2-a4c9-28f9a5ddd044","Type":"ContainerDied","Data":"8b7241b73e8a1ff5b13e5732d0d1c9ab20a45823a3e5454d8a97537de5a9680c"} Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.453747 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" event={"ID":"00d36531-cc42-49a2-a4c9-28f9a5ddd044","Type":"ContainerDied","Data":"adff7418b9342665f8a003a2f969034eec7663da81d7dda35af281f14d4f5cda"} Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.453768 4894 scope.go:117] "RemoveContainer" containerID="8b7241b73e8a1ff5b13e5732d0d1c9ab20a45823a3e5454d8a97537de5a9680c" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.453894 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-qjlt6" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.455875 4894 generic.go:334] "Generic (PLEG): container finished" podID="171c863a-36f5-48c0-be81-752684537f0b" containerID="79966100b093d6f879432f696bf6008f04634486cdb672b037d179f09313946f" exitCode=0 Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.455906 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" event={"ID":"171c863a-36f5-48c0-be81-752684537f0b","Type":"ContainerDied","Data":"79966100b093d6f879432f696bf6008f04634486cdb672b037d179f09313946f"} Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.455924 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" event={"ID":"171c863a-36f5-48c0-be81-752684537f0b","Type":"ContainerDied","Data":"0e757050ff3deaa03ba4dde7fb53c2f722babb0539f85f88aee92a279f5ec20c"} Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.455961 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.463565 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/00d36531-cc42-49a2-a4c9-28f9a5ddd044-proxy-ca-bundles\") pod \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\" (UID: \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\") " Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.463610 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqfdg\" (UniqueName: \"kubernetes.io/projected/00d36531-cc42-49a2-a4c9-28f9a5ddd044-kube-api-access-qqfdg\") pod \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\" (UID: \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\") " Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.463653 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-client-ca\") pod \"171c863a-36f5-48c0-be81-752684537f0b\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.463721 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/171c863a-36f5-48c0-be81-752684537f0b-serving-cert\") pod \"171c863a-36f5-48c0-be81-752684537f0b\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.463747 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00d36531-cc42-49a2-a4c9-28f9a5ddd044-config\") pod \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\" (UID: \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\") " Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.463771 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9sjj\" (UniqueName: \"kubernetes.io/projected/171c863a-36f5-48c0-be81-752684537f0b-kube-api-access-h9sjj\") pod \"171c863a-36f5-48c0-be81-752684537f0b\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.463791 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-config\") pod \"171c863a-36f5-48c0-be81-752684537f0b\" (UID: \"171c863a-36f5-48c0-be81-752684537f0b\") " Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.463814 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/00d36531-cc42-49a2-a4c9-28f9a5ddd044-client-ca\") pod \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\" (UID: \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\") " Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.463838 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00d36531-cc42-49a2-a4c9-28f9a5ddd044-serving-cert\") pod \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\" (UID: \"00d36531-cc42-49a2-a4c9-28f9a5ddd044\") " Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.464465 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00d36531-cc42-49a2-a4c9-28f9a5ddd044-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "00d36531-cc42-49a2-a4c9-28f9a5ddd044" (UID: "00d36531-cc42-49a2-a4c9-28f9a5ddd044"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.465015 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00d36531-cc42-49a2-a4c9-28f9a5ddd044-client-ca" (OuterVolumeSpecName: "client-ca") pod "00d36531-cc42-49a2-a4c9-28f9a5ddd044" (UID: "00d36531-cc42-49a2-a4c9-28f9a5ddd044"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.465172 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-client-ca" (OuterVolumeSpecName: "client-ca") pod "171c863a-36f5-48c0-be81-752684537f0b" (UID: "171c863a-36f5-48c0-be81-752684537f0b"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.466061 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/00d36531-cc42-49a2-a4c9-28f9a5ddd044-config" (OuterVolumeSpecName: "config") pod "00d36531-cc42-49a2-a4c9-28f9a5ddd044" (UID: "00d36531-cc42-49a2-a4c9-28f9a5ddd044"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.466867 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-config" (OuterVolumeSpecName: "config") pod "171c863a-36f5-48c0-be81-752684537f0b" (UID: "171c863a-36f5-48c0-be81-752684537f0b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.470061 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/171c863a-36f5-48c0-be81-752684537f0b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "171c863a-36f5-48c0-be81-752684537f0b" (UID: "171c863a-36f5-48c0-be81-752684537f0b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.470202 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/171c863a-36f5-48c0-be81-752684537f0b-kube-api-access-h9sjj" (OuterVolumeSpecName: "kube-api-access-h9sjj") pod "171c863a-36f5-48c0-be81-752684537f0b" (UID: "171c863a-36f5-48c0-be81-752684537f0b"). InnerVolumeSpecName "kube-api-access-h9sjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.476344 4894 scope.go:117] "RemoveContainer" containerID="8b7241b73e8a1ff5b13e5732d0d1c9ab20a45823a3e5454d8a97537de5a9680c" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.477360 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/00d36531-cc42-49a2-a4c9-28f9a5ddd044-kube-api-access-qqfdg" (OuterVolumeSpecName: "kube-api-access-qqfdg") pod "00d36531-cc42-49a2-a4c9-28f9a5ddd044" (UID: "00d36531-cc42-49a2-a4c9-28f9a5ddd044"). InnerVolumeSpecName "kube-api-access-qqfdg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:36:47 crc kubenswrapper[4894]: E1209 15:36:47.477754 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b7241b73e8a1ff5b13e5732d0d1c9ab20a45823a3e5454d8a97537de5a9680c\": container with ID starting with 8b7241b73e8a1ff5b13e5732d0d1c9ab20a45823a3e5454d8a97537de5a9680c not found: ID does not exist" containerID="8b7241b73e8a1ff5b13e5732d0d1c9ab20a45823a3e5454d8a97537de5a9680c" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.477819 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b7241b73e8a1ff5b13e5732d0d1c9ab20a45823a3e5454d8a97537de5a9680c"} err="failed to get container status \"8b7241b73e8a1ff5b13e5732d0d1c9ab20a45823a3e5454d8a97537de5a9680c\": rpc error: code = NotFound desc = could not find container \"8b7241b73e8a1ff5b13e5732d0d1c9ab20a45823a3e5454d8a97537de5a9680c\": container with ID starting with 8b7241b73e8a1ff5b13e5732d0d1c9ab20a45823a3e5454d8a97537de5a9680c not found: ID does not exist" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.477841 4894 scope.go:117] "RemoveContainer" containerID="79966100b093d6f879432f696bf6008f04634486cdb672b037d179f09313946f" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.479217 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/00d36531-cc42-49a2-a4c9-28f9a5ddd044-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "00d36531-cc42-49a2-a4c9-28f9a5ddd044" (UID: "00d36531-cc42-49a2-a4c9-28f9a5ddd044"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.491499 4894 scope.go:117] "RemoveContainer" containerID="79966100b093d6f879432f696bf6008f04634486cdb672b037d179f09313946f" Dec 09 15:36:47 crc kubenswrapper[4894]: E1209 15:36:47.491854 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"79966100b093d6f879432f696bf6008f04634486cdb672b037d179f09313946f\": container with ID starting with 79966100b093d6f879432f696bf6008f04634486cdb672b037d179f09313946f not found: ID does not exist" containerID="79966100b093d6f879432f696bf6008f04634486cdb672b037d179f09313946f" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.491885 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"79966100b093d6f879432f696bf6008f04634486cdb672b037d179f09313946f"} err="failed to get container status \"79966100b093d6f879432f696bf6008f04634486cdb672b037d179f09313946f\": rpc error: code = NotFound desc = could not find container \"79966100b093d6f879432f696bf6008f04634486cdb672b037d179f09313946f\": container with ID starting with 79966100b093d6f879432f696bf6008f04634486cdb672b037d179f09313946f not found: ID does not exist" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.565385 4894 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/00d36531-cc42-49a2-a4c9-28f9a5ddd044-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.565419 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqfdg\" (UniqueName: \"kubernetes.io/projected/00d36531-cc42-49a2-a4c9-28f9a5ddd044-kube-api-access-qqfdg\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.565431 4894 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.565441 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/171c863a-36f5-48c0-be81-752684537f0b-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.565452 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/00d36531-cc42-49a2-a4c9-28f9a5ddd044-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.565464 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9sjj\" (UniqueName: \"kubernetes.io/projected/171c863a-36f5-48c0-be81-752684537f0b-kube-api-access-h9sjj\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.565475 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/171c863a-36f5-48c0-be81-752684537f0b-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.565487 4894 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/00d36531-cc42-49a2-a4c9-28f9a5ddd044-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.565497 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/00d36531-cc42-49a2-a4c9-28f9a5ddd044-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.783588 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826"] Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.787166 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-56826"] Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.796085 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qjlt6"] Dec 09 15:36:47 crc kubenswrapper[4894]: I1209 15:36:47.798532 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-qjlt6"] Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.112598 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="00d36531-cc42-49a2-a4c9-28f9a5ddd044" path="/var/lib/kubelet/pods/00d36531-cc42-49a2-a4c9-28f9a5ddd044/volumes" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.113199 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="171c863a-36f5-48c0-be81-752684537f0b" path="/var/lib/kubelet/pods/171c863a-36f5-48c0-be81-752684537f0b/volumes" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.447546 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp"] Dec 09 15:36:48 crc kubenswrapper[4894]: E1209 15:36:48.447771 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c278157d-d806-4fa1-a597-076d95a286e2" containerName="extract-utilities" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.447784 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c278157d-d806-4fa1-a597-076d95a286e2" containerName="extract-utilities" Dec 09 15:36:48 crc kubenswrapper[4894]: E1209 15:36:48.447793 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="171c863a-36f5-48c0-be81-752684537f0b" containerName="route-controller-manager" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.447800 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="171c863a-36f5-48c0-be81-752684537f0b" containerName="route-controller-manager" Dec 09 15:36:48 crc kubenswrapper[4894]: E1209 15:36:48.447807 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" containerName="registry-server" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.447812 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" containerName="registry-server" Dec 09 15:36:48 crc kubenswrapper[4894]: E1209 15:36:48.447823 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c278157d-d806-4fa1-a597-076d95a286e2" containerName="extract-content" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.447828 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c278157d-d806-4fa1-a597-076d95a286e2" containerName="extract-content" Dec 09 15:36:48 crc kubenswrapper[4894]: E1209 15:36:48.447836 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" containerName="extract-content" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.447842 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" containerName="extract-content" Dec 09 15:36:48 crc kubenswrapper[4894]: E1209 15:36:48.447851 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4397f08d-278a-4e2d-9664-e44161c96638" containerName="extract-content" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.447857 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="4397f08d-278a-4e2d-9664-e44161c96638" containerName="extract-content" Dec 09 15:36:48 crc kubenswrapper[4894]: E1209 15:36:48.447869 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="00d36531-cc42-49a2-a4c9-28f9a5ddd044" containerName="controller-manager" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.447876 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="00d36531-cc42-49a2-a4c9-28f9a5ddd044" containerName="controller-manager" Dec 09 15:36:48 crc kubenswrapper[4894]: E1209 15:36:48.447884 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" containerName="extract-utilities" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.447891 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" containerName="extract-utilities" Dec 09 15:36:48 crc kubenswrapper[4894]: E1209 15:36:48.447898 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" containerName="registry-server" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.447903 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" containerName="registry-server" Dec 09 15:36:48 crc kubenswrapper[4894]: E1209 15:36:48.447910 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2421584b-8a4b-4cb1-97c0-cb41d8ec16b2" containerName="extract-content" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.447917 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2421584b-8a4b-4cb1-97c0-cb41d8ec16b2" containerName="extract-content" Dec 09 15:36:48 crc kubenswrapper[4894]: E1209 15:36:48.447926 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0847bb67-0cb7-4616-b715-29b6f12f67b5" containerName="marketplace-operator" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.447931 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0847bb67-0cb7-4616-b715-29b6f12f67b5" containerName="marketplace-operator" Dec 09 15:36:48 crc kubenswrapper[4894]: E1209 15:36:48.447941 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2421584b-8a4b-4cb1-97c0-cb41d8ec16b2" containerName="extract-utilities" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.447947 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2421584b-8a4b-4cb1-97c0-cb41d8ec16b2" containerName="extract-utilities" Dec 09 15:36:48 crc kubenswrapper[4894]: E1209 15:36:48.447956 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c278157d-d806-4fa1-a597-076d95a286e2" containerName="registry-server" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.447961 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c278157d-d806-4fa1-a597-076d95a286e2" containerName="registry-server" Dec 09 15:36:48 crc kubenswrapper[4894]: E1209 15:36:48.447967 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" containerName="extract-content" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.447972 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" containerName="extract-content" Dec 09 15:36:48 crc kubenswrapper[4894]: E1209 15:36:48.447980 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" containerName="extract-utilities" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.447985 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" containerName="extract-utilities" Dec 09 15:36:48 crc kubenswrapper[4894]: E1209 15:36:48.447995 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4397f08d-278a-4e2d-9664-e44161c96638" containerName="extract-utilities" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.448001 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="4397f08d-278a-4e2d-9664-e44161c96638" containerName="extract-utilities" Dec 09 15:36:48 crc kubenswrapper[4894]: E1209 15:36:48.448008 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2421584b-8a4b-4cb1-97c0-cb41d8ec16b2" containerName="registry-server" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.448014 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2421584b-8a4b-4cb1-97c0-cb41d8ec16b2" containerName="registry-server" Dec 09 15:36:48 crc kubenswrapper[4894]: E1209 15:36:48.448022 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4397f08d-278a-4e2d-9664-e44161c96638" containerName="registry-server" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.448028 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="4397f08d-278a-4e2d-9664-e44161c96638" containerName="registry-server" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.448139 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee374eee-998a-48f2-acdf-39daa5ce2fdd" containerName="registry-server" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.448150 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="c278157d-d806-4fa1-a597-076d95a286e2" containerName="registry-server" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.448159 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0847bb67-0cb7-4616-b715-29b6f12f67b5" containerName="marketplace-operator" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.448167 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="4397f08d-278a-4e2d-9664-e44161c96638" containerName="registry-server" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.448176 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfb801bd-e785-4fdc-8a34-d2b8769d3a29" containerName="registry-server" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.448187 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="2421584b-8a4b-4cb1-97c0-cb41d8ec16b2" containerName="registry-server" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.448201 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="171c863a-36f5-48c0-be81-752684537f0b" containerName="route-controller-manager" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.448212 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="00d36531-cc42-49a2-a4c9-28f9a5ddd044" containerName="controller-manager" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.448550 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.450806 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.451078 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.452443 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.452468 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.452665 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.452713 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.456047 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5b6d94b98d-mf46v"] Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.456887 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.467199 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.467363 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.467410 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.467546 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.469219 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.469230 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.473022 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.485725 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5b6d94b98d-mf46v"] Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.498205 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp"] Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.576596 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-client-ca\") pod \"route-controller-manager-757b9d8c6-m66jp\" (UID: \"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669\") " pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.576922 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6st4f\" (UniqueName: \"kubernetes.io/projected/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-kube-api-access-6st4f\") pod \"route-controller-manager-757b9d8c6-m66jp\" (UID: \"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669\") " pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.576954 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/68ebb693-9d20-4157-b90c-c827e52a1ea4-client-ca\") pod \"controller-manager-5b6d94b98d-mf46v\" (UID: \"68ebb693-9d20-4157-b90c-c827e52a1ea4\") " pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.576989 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68ebb693-9d20-4157-b90c-c827e52a1ea4-serving-cert\") pod \"controller-manager-5b6d94b98d-mf46v\" (UID: \"68ebb693-9d20-4157-b90c-c827e52a1ea4\") " pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.577014 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68ebb693-9d20-4157-b90c-c827e52a1ea4-config\") pod \"controller-manager-5b6d94b98d-mf46v\" (UID: \"68ebb693-9d20-4157-b90c-c827e52a1ea4\") " pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.577086 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-config\") pod \"route-controller-manager-757b9d8c6-m66jp\" (UID: \"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669\") " pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.577109 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/68ebb693-9d20-4157-b90c-c827e52a1ea4-proxy-ca-bundles\") pod \"controller-manager-5b6d94b98d-mf46v\" (UID: \"68ebb693-9d20-4157-b90c-c827e52a1ea4\") " pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.577127 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss66b\" (UniqueName: \"kubernetes.io/projected/68ebb693-9d20-4157-b90c-c827e52a1ea4-kube-api-access-ss66b\") pod \"controller-manager-5b6d94b98d-mf46v\" (UID: \"68ebb693-9d20-4157-b90c-c827e52a1ea4\") " pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.577147 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-serving-cert\") pod \"route-controller-manager-757b9d8c6-m66jp\" (UID: \"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669\") " pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.677509 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68ebb693-9d20-4157-b90c-c827e52a1ea4-config\") pod \"controller-manager-5b6d94b98d-mf46v\" (UID: \"68ebb693-9d20-4157-b90c-c827e52a1ea4\") " pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.677566 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-config\") pod \"route-controller-manager-757b9d8c6-m66jp\" (UID: \"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669\") " pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.677582 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/68ebb693-9d20-4157-b90c-c827e52a1ea4-proxy-ca-bundles\") pod \"controller-manager-5b6d94b98d-mf46v\" (UID: \"68ebb693-9d20-4157-b90c-c827e52a1ea4\") " pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.677601 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ss66b\" (UniqueName: \"kubernetes.io/projected/68ebb693-9d20-4157-b90c-c827e52a1ea4-kube-api-access-ss66b\") pod \"controller-manager-5b6d94b98d-mf46v\" (UID: \"68ebb693-9d20-4157-b90c-c827e52a1ea4\") " pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.677620 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-serving-cert\") pod \"route-controller-manager-757b9d8c6-m66jp\" (UID: \"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669\") " pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.677657 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-client-ca\") pod \"route-controller-manager-757b9d8c6-m66jp\" (UID: \"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669\") " pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.677682 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6st4f\" (UniqueName: \"kubernetes.io/projected/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-kube-api-access-6st4f\") pod \"route-controller-manager-757b9d8c6-m66jp\" (UID: \"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669\") " pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.678359 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/68ebb693-9d20-4157-b90c-c827e52a1ea4-client-ca\") pod \"controller-manager-5b6d94b98d-mf46v\" (UID: \"68ebb693-9d20-4157-b90c-c827e52a1ea4\") " pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.678488 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68ebb693-9d20-4157-b90c-c827e52a1ea4-serving-cert\") pod \"controller-manager-5b6d94b98d-mf46v\" (UID: \"68ebb693-9d20-4157-b90c-c827e52a1ea4\") " pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.679383 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/68ebb693-9d20-4157-b90c-c827e52a1ea4-client-ca\") pod \"controller-manager-5b6d94b98d-mf46v\" (UID: \"68ebb693-9d20-4157-b90c-c827e52a1ea4\") " pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.679448 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-client-ca\") pod \"route-controller-manager-757b9d8c6-m66jp\" (UID: \"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669\") " pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.679483 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-config\") pod \"route-controller-manager-757b9d8c6-m66jp\" (UID: \"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669\") " pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.679771 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68ebb693-9d20-4157-b90c-c827e52a1ea4-config\") pod \"controller-manager-5b6d94b98d-mf46v\" (UID: \"68ebb693-9d20-4157-b90c-c827e52a1ea4\") " pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.679927 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/68ebb693-9d20-4157-b90c-c827e52a1ea4-proxy-ca-bundles\") pod \"controller-manager-5b6d94b98d-mf46v\" (UID: \"68ebb693-9d20-4157-b90c-c827e52a1ea4\") " pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.683968 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-serving-cert\") pod \"route-controller-manager-757b9d8c6-m66jp\" (UID: \"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669\") " pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.690593 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68ebb693-9d20-4157-b90c-c827e52a1ea4-serving-cert\") pod \"controller-manager-5b6d94b98d-mf46v\" (UID: \"68ebb693-9d20-4157-b90c-c827e52a1ea4\") " pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.693409 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ss66b\" (UniqueName: \"kubernetes.io/projected/68ebb693-9d20-4157-b90c-c827e52a1ea4-kube-api-access-ss66b\") pod \"controller-manager-5b6d94b98d-mf46v\" (UID: \"68ebb693-9d20-4157-b90c-c827e52a1ea4\") " pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.701882 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6st4f\" (UniqueName: \"kubernetes.io/projected/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-kube-api-access-6st4f\") pod \"route-controller-manager-757b9d8c6-m66jp\" (UID: \"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669\") " pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.772853 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.788628 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.957890 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp"] Dec 09 15:36:48 crc kubenswrapper[4894]: I1209 15:36:48.995985 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5b6d94b98d-mf46v"] Dec 09 15:36:49 crc kubenswrapper[4894]: W1209 15:36:49.002949 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68ebb693_9d20_4157_b90c_c827e52a1ea4.slice/crio-8c8b81a97a77dcb2415fb9343e646b49b5e7b1c09a4ec919ee251494d7fb9b4c WatchSource:0}: Error finding container 8c8b81a97a77dcb2415fb9343e646b49b5e7b1c09a4ec919ee251494d7fb9b4c: Status 404 returned error can't find the container with id 8c8b81a97a77dcb2415fb9343e646b49b5e7b1c09a4ec919ee251494d7fb9b4c Dec 09 15:36:49 crc kubenswrapper[4894]: I1209 15:36:49.478523 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" event={"ID":"68ebb693-9d20-4157-b90c-c827e52a1ea4","Type":"ContainerStarted","Data":"8c8b81a97a77dcb2415fb9343e646b49b5e7b1c09a4ec919ee251494d7fb9b4c"} Dec 09 15:36:49 crc kubenswrapper[4894]: I1209 15:36:49.479657 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" event={"ID":"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669","Type":"ContainerStarted","Data":"98ad427223716f2ac796bf572d1b69d503fcec485c49aacf61ac2ae628c1f2d0"} Dec 09 15:36:50 crc kubenswrapper[4894]: I1209 15:36:50.490476 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" event={"ID":"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669","Type":"ContainerStarted","Data":"c1c316251ce54444d556419d520ea04f60d388f204d65c7fd58fcc86f9e2c2e7"} Dec 09 15:36:50 crc kubenswrapper[4894]: I1209 15:36:50.490554 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" Dec 09 15:36:50 crc kubenswrapper[4894]: I1209 15:36:50.491923 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" event={"ID":"68ebb693-9d20-4157-b90c-c827e52a1ea4","Type":"ContainerStarted","Data":"a4abcfdb1a2c5536dd07b89e7ba036e472216478c386587abacb764eafda6fd1"} Dec 09 15:36:50 crc kubenswrapper[4894]: I1209 15:36:50.492308 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" Dec 09 15:36:50 crc kubenswrapper[4894]: I1209 15:36:50.496323 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" Dec 09 15:36:50 crc kubenswrapper[4894]: I1209 15:36:50.528268 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" podStartSLOduration=3.5282473960000003 podStartE2EDuration="3.528247396s" podCreationTimestamp="2025-12-09 15:36:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:36:50.510703574 +0000 UTC m=+304.829914253" watchObservedRunningTime="2025-12-09 15:36:50.528247396 +0000 UTC m=+304.847458075" Dec 09 15:36:50 crc kubenswrapper[4894]: I1209 15:36:50.899043 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" Dec 09 15:36:50 crc kubenswrapper[4894]: I1209 15:36:50.918060 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" podStartSLOduration=3.918045158 podStartE2EDuration="3.918045158s" podCreationTimestamp="2025-12-09 15:36:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:36:50.52802451 +0000 UTC m=+304.847235179" watchObservedRunningTime="2025-12-09 15:36:50.918045158 +0000 UTC m=+305.237255817" Dec 09 15:36:52 crc kubenswrapper[4894]: I1209 15:36:52.584197 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5b6d94b98d-mf46v"] Dec 09 15:36:52 crc kubenswrapper[4894]: I1209 15:36:52.613011 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp"] Dec 09 15:36:53 crc kubenswrapper[4894]: I1209 15:36:53.507374 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" podUID="68ebb693-9d20-4157-b90c-c827e52a1ea4" containerName="controller-manager" containerID="cri-o://a4abcfdb1a2c5536dd07b89e7ba036e472216478c386587abacb764eafda6fd1" gracePeriod=30 Dec 09 15:36:53 crc kubenswrapper[4894]: I1209 15:36:53.507292 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" podUID="cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669" containerName="route-controller-manager" containerID="cri-o://c1c316251ce54444d556419d520ea04f60d388f204d65c7fd58fcc86f9e2c2e7" gracePeriod=30 Dec 09 15:36:54 crc kubenswrapper[4894]: I1209 15:36:54.512839 4894 generic.go:334] "Generic (PLEG): container finished" podID="68ebb693-9d20-4157-b90c-c827e52a1ea4" containerID="a4abcfdb1a2c5536dd07b89e7ba036e472216478c386587abacb764eafda6fd1" exitCode=0 Dec 09 15:36:54 crc kubenswrapper[4894]: I1209 15:36:54.512961 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" event={"ID":"68ebb693-9d20-4157-b90c-c827e52a1ea4","Type":"ContainerDied","Data":"a4abcfdb1a2c5536dd07b89e7ba036e472216478c386587abacb764eafda6fd1"} Dec 09 15:36:54 crc kubenswrapper[4894]: I1209 15:36:54.515789 4894 generic.go:334] "Generic (PLEG): container finished" podID="cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669" containerID="c1c316251ce54444d556419d520ea04f60d388f204d65c7fd58fcc86f9e2c2e7" exitCode=0 Dec 09 15:36:54 crc kubenswrapper[4894]: I1209 15:36:54.515878 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" event={"ID":"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669","Type":"ContainerDied","Data":"c1c316251ce54444d556419d520ea04f60d388f204d65c7fd58fcc86f9e2c2e7"} Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.051254 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.108166 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.154101 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6st4f\" (UniqueName: \"kubernetes.io/projected/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-kube-api-access-6st4f\") pod \"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669\" (UID: \"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669\") " Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.154162 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-serving-cert\") pod \"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669\" (UID: \"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669\") " Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.154227 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-client-ca\") pod \"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669\" (UID: \"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669\") " Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.154270 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-config\") pod \"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669\" (UID: \"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669\") " Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.155123 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-config" (OuterVolumeSpecName: "config") pod "cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669" (UID: "cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.155515 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-client-ca" (OuterVolumeSpecName: "client-ca") pod "cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669" (UID: "cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.160387 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669" (UID: "cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.161859 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-kube-api-access-6st4f" (OuterVolumeSpecName: "kube-api-access-6st4f") pod "cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669" (UID: "cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669"). InnerVolumeSpecName "kube-api-access-6st4f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.255910 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68ebb693-9d20-4157-b90c-c827e52a1ea4-serving-cert\") pod \"68ebb693-9d20-4157-b90c-c827e52a1ea4\" (UID: \"68ebb693-9d20-4157-b90c-c827e52a1ea4\") " Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.256004 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/68ebb693-9d20-4157-b90c-c827e52a1ea4-client-ca\") pod \"68ebb693-9d20-4157-b90c-c827e52a1ea4\" (UID: \"68ebb693-9d20-4157-b90c-c827e52a1ea4\") " Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.256032 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ss66b\" (UniqueName: \"kubernetes.io/projected/68ebb693-9d20-4157-b90c-c827e52a1ea4-kube-api-access-ss66b\") pod \"68ebb693-9d20-4157-b90c-c827e52a1ea4\" (UID: \"68ebb693-9d20-4157-b90c-c827e52a1ea4\") " Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.256109 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68ebb693-9d20-4157-b90c-c827e52a1ea4-config\") pod \"68ebb693-9d20-4157-b90c-c827e52a1ea4\" (UID: \"68ebb693-9d20-4157-b90c-c827e52a1ea4\") " Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.256139 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/68ebb693-9d20-4157-b90c-c827e52a1ea4-proxy-ca-bundles\") pod \"68ebb693-9d20-4157-b90c-c827e52a1ea4\" (UID: \"68ebb693-9d20-4157-b90c-c827e52a1ea4\") " Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.256392 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6st4f\" (UniqueName: \"kubernetes.io/projected/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-kube-api-access-6st4f\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.256407 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.256420 4894 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.256432 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.256683 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68ebb693-9d20-4157-b90c-c827e52a1ea4-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "68ebb693-9d20-4157-b90c-c827e52a1ea4" (UID: "68ebb693-9d20-4157-b90c-c827e52a1ea4"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.256692 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68ebb693-9d20-4157-b90c-c827e52a1ea4-config" (OuterVolumeSpecName: "config") pod "68ebb693-9d20-4157-b90c-c827e52a1ea4" (UID: "68ebb693-9d20-4157-b90c-c827e52a1ea4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.256997 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/68ebb693-9d20-4157-b90c-c827e52a1ea4-client-ca" (OuterVolumeSpecName: "client-ca") pod "68ebb693-9d20-4157-b90c-c827e52a1ea4" (UID: "68ebb693-9d20-4157-b90c-c827e52a1ea4"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.258895 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/68ebb693-9d20-4157-b90c-c827e52a1ea4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "68ebb693-9d20-4157-b90c-c827e52a1ea4" (UID: "68ebb693-9d20-4157-b90c-c827e52a1ea4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.259501 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68ebb693-9d20-4157-b90c-c827e52a1ea4-kube-api-access-ss66b" (OuterVolumeSpecName: "kube-api-access-ss66b") pod "68ebb693-9d20-4157-b90c-c827e52a1ea4" (UID: "68ebb693-9d20-4157-b90c-c827e52a1ea4"). InnerVolumeSpecName "kube-api-access-ss66b". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.357511 4894 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/68ebb693-9d20-4157-b90c-c827e52a1ea4-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.357559 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/68ebb693-9d20-4157-b90c-c827e52a1ea4-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.357572 4894 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/68ebb693-9d20-4157-b90c-c827e52a1ea4-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.357588 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ss66b\" (UniqueName: \"kubernetes.io/projected/68ebb693-9d20-4157-b90c-c827e52a1ea4-kube-api-access-ss66b\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.357600 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/68ebb693-9d20-4157-b90c-c827e52a1ea4-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.453357 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd"] Dec 09 15:36:55 crc kubenswrapper[4894]: E1209 15:36:55.453587 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68ebb693-9d20-4157-b90c-c827e52a1ea4" containerName="controller-manager" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.453605 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="68ebb693-9d20-4157-b90c-c827e52a1ea4" containerName="controller-manager" Dec 09 15:36:55 crc kubenswrapper[4894]: E1209 15:36:55.453623 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669" containerName="route-controller-manager" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.453671 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669" containerName="route-controller-manager" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.453778 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669" containerName="route-controller-manager" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.453802 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="68ebb693-9d20-4157-b90c-c827e52a1ea4" containerName="controller-manager" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.454227 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.464966 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd"] Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.522352 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" event={"ID":"68ebb693-9d20-4157-b90c-c827e52a1ea4","Type":"ContainerDied","Data":"8c8b81a97a77dcb2415fb9343e646b49b5e7b1c09a4ec919ee251494d7fb9b4c"} Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.522409 4894 scope.go:117] "RemoveContainer" containerID="a4abcfdb1a2c5536dd07b89e7ba036e472216478c386587abacb764eafda6fd1" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.522529 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5b6d94b98d-mf46v" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.525400 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" event={"ID":"cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669","Type":"ContainerDied","Data":"98ad427223716f2ac796bf572d1b69d503fcec485c49aacf61ac2ae628c1f2d0"} Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.525507 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.541029 4894 scope.go:117] "RemoveContainer" containerID="c1c316251ce54444d556419d520ea04f60d388f204d65c7fd58fcc86f9e2c2e7" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.559622 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/209ad566-6ec9-44ba-9a0a-bc62cbfea583-client-ca\") pod \"route-controller-manager-7484d9ddcc-44hvd\" (UID: \"209ad566-6ec9-44ba-9a0a-bc62cbfea583\") " pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.559936 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf67l\" (UniqueName: \"kubernetes.io/projected/209ad566-6ec9-44ba-9a0a-bc62cbfea583-kube-api-access-qf67l\") pod \"route-controller-manager-7484d9ddcc-44hvd\" (UID: \"209ad566-6ec9-44ba-9a0a-bc62cbfea583\") " pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.560077 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/209ad566-6ec9-44ba-9a0a-bc62cbfea583-config\") pod \"route-controller-manager-7484d9ddcc-44hvd\" (UID: \"209ad566-6ec9-44ba-9a0a-bc62cbfea583\") " pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.560179 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/209ad566-6ec9-44ba-9a0a-bc62cbfea583-serving-cert\") pod \"route-controller-manager-7484d9ddcc-44hvd\" (UID: \"209ad566-6ec9-44ba-9a0a-bc62cbfea583\") " pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.560351 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-5b6d94b98d-mf46v"] Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.563996 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-5b6d94b98d-mf46v"] Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.575060 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp"] Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.578369 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-757b9d8c6-m66jp"] Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.661412 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/209ad566-6ec9-44ba-9a0a-bc62cbfea583-client-ca\") pod \"route-controller-manager-7484d9ddcc-44hvd\" (UID: \"209ad566-6ec9-44ba-9a0a-bc62cbfea583\") " pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.661472 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf67l\" (UniqueName: \"kubernetes.io/projected/209ad566-6ec9-44ba-9a0a-bc62cbfea583-kube-api-access-qf67l\") pod \"route-controller-manager-7484d9ddcc-44hvd\" (UID: \"209ad566-6ec9-44ba-9a0a-bc62cbfea583\") " pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.661504 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/209ad566-6ec9-44ba-9a0a-bc62cbfea583-config\") pod \"route-controller-manager-7484d9ddcc-44hvd\" (UID: \"209ad566-6ec9-44ba-9a0a-bc62cbfea583\") " pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.661529 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/209ad566-6ec9-44ba-9a0a-bc62cbfea583-serving-cert\") pod \"route-controller-manager-7484d9ddcc-44hvd\" (UID: \"209ad566-6ec9-44ba-9a0a-bc62cbfea583\") " pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.662496 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/209ad566-6ec9-44ba-9a0a-bc62cbfea583-client-ca\") pod \"route-controller-manager-7484d9ddcc-44hvd\" (UID: \"209ad566-6ec9-44ba-9a0a-bc62cbfea583\") " pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.663008 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/209ad566-6ec9-44ba-9a0a-bc62cbfea583-config\") pod \"route-controller-manager-7484d9ddcc-44hvd\" (UID: \"209ad566-6ec9-44ba-9a0a-bc62cbfea583\") " pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.666482 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/209ad566-6ec9-44ba-9a0a-bc62cbfea583-serving-cert\") pod \"route-controller-manager-7484d9ddcc-44hvd\" (UID: \"209ad566-6ec9-44ba-9a0a-bc62cbfea583\") " pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.676027 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf67l\" (UniqueName: \"kubernetes.io/projected/209ad566-6ec9-44ba-9a0a-bc62cbfea583-kube-api-access-qf67l\") pod \"route-controller-manager-7484d9ddcc-44hvd\" (UID: \"209ad566-6ec9-44ba-9a0a-bc62cbfea583\") " pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.781478 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" Dec 09 15:36:55 crc kubenswrapper[4894]: I1209 15:36:55.953793 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd"] Dec 09 15:36:56 crc kubenswrapper[4894]: I1209 15:36:56.112753 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68ebb693-9d20-4157-b90c-c827e52a1ea4" path="/var/lib/kubelet/pods/68ebb693-9d20-4157-b90c-c827e52a1ea4/volumes" Dec 09 15:36:56 crc kubenswrapper[4894]: I1209 15:36:56.114225 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669" path="/var/lib/kubelet/pods/cbd8fd13-aa17-4b6d-9d5e-dada3ed6d669/volumes" Dec 09 15:36:56 crc kubenswrapper[4894]: I1209 15:36:56.533283 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" event={"ID":"209ad566-6ec9-44ba-9a0a-bc62cbfea583","Type":"ContainerStarted","Data":"6a5a84367ea89e1aa00cd610ed60fa492af84bf3b3be6523d156055c59749b1a"} Dec 09 15:36:56 crc kubenswrapper[4894]: I1209 15:36:56.533334 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" event={"ID":"209ad566-6ec9-44ba-9a0a-bc62cbfea583","Type":"ContainerStarted","Data":"f7ab22717de4904698f011724e58dc882364726f4bdedc43e02adfa1684231a2"} Dec 09 15:36:56 crc kubenswrapper[4894]: I1209 15:36:56.533542 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" Dec 09 15:36:56 crc kubenswrapper[4894]: I1209 15:36:56.551925 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" podStartSLOduration=2.5519101749999997 podStartE2EDuration="2.551910175s" podCreationTimestamp="2025-12-09 15:36:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:36:56.550689482 +0000 UTC m=+310.869900151" watchObservedRunningTime="2025-12-09 15:36:56.551910175 +0000 UTC m=+310.871120844" Dec 09 15:36:56 crc kubenswrapper[4894]: I1209 15:36:56.739184 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.452095 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-6bb7b6db47-srl48"] Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.452836 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bb7b6db47-srl48" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.454879 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.454924 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.455247 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.455269 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.455396 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.456143 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.468280 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.472151 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6bb7b6db47-srl48"] Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.585909 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0deb5e73-3948-40cb-8dca-c1ebcf8c117f-client-ca\") pod \"controller-manager-6bb7b6db47-srl48\" (UID: \"0deb5e73-3948-40cb-8dca-c1ebcf8c117f\") " pod="openshift-controller-manager/controller-manager-6bb7b6db47-srl48" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.586330 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f955r\" (UniqueName: \"kubernetes.io/projected/0deb5e73-3948-40cb-8dca-c1ebcf8c117f-kube-api-access-f955r\") pod \"controller-manager-6bb7b6db47-srl48\" (UID: \"0deb5e73-3948-40cb-8dca-c1ebcf8c117f\") " pod="openshift-controller-manager/controller-manager-6bb7b6db47-srl48" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.586397 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0deb5e73-3948-40cb-8dca-c1ebcf8c117f-config\") pod \"controller-manager-6bb7b6db47-srl48\" (UID: \"0deb5e73-3948-40cb-8dca-c1ebcf8c117f\") " pod="openshift-controller-manager/controller-manager-6bb7b6db47-srl48" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.586496 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0deb5e73-3948-40cb-8dca-c1ebcf8c117f-serving-cert\") pod \"controller-manager-6bb7b6db47-srl48\" (UID: \"0deb5e73-3948-40cb-8dca-c1ebcf8c117f\") " pod="openshift-controller-manager/controller-manager-6bb7b6db47-srl48" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.586522 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0deb5e73-3948-40cb-8dca-c1ebcf8c117f-proxy-ca-bundles\") pod \"controller-manager-6bb7b6db47-srl48\" (UID: \"0deb5e73-3948-40cb-8dca-c1ebcf8c117f\") " pod="openshift-controller-manager/controller-manager-6bb7b6db47-srl48" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.688001 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0deb5e73-3948-40cb-8dca-c1ebcf8c117f-client-ca\") pod \"controller-manager-6bb7b6db47-srl48\" (UID: \"0deb5e73-3948-40cb-8dca-c1ebcf8c117f\") " pod="openshift-controller-manager/controller-manager-6bb7b6db47-srl48" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.688103 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f955r\" (UniqueName: \"kubernetes.io/projected/0deb5e73-3948-40cb-8dca-c1ebcf8c117f-kube-api-access-f955r\") pod \"controller-manager-6bb7b6db47-srl48\" (UID: \"0deb5e73-3948-40cb-8dca-c1ebcf8c117f\") " pod="openshift-controller-manager/controller-manager-6bb7b6db47-srl48" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.688148 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0deb5e73-3948-40cb-8dca-c1ebcf8c117f-config\") pod \"controller-manager-6bb7b6db47-srl48\" (UID: \"0deb5e73-3948-40cb-8dca-c1ebcf8c117f\") " pod="openshift-controller-manager/controller-manager-6bb7b6db47-srl48" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.688194 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0deb5e73-3948-40cb-8dca-c1ebcf8c117f-serving-cert\") pod \"controller-manager-6bb7b6db47-srl48\" (UID: \"0deb5e73-3948-40cb-8dca-c1ebcf8c117f\") " pod="openshift-controller-manager/controller-manager-6bb7b6db47-srl48" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.688217 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0deb5e73-3948-40cb-8dca-c1ebcf8c117f-proxy-ca-bundles\") pod \"controller-manager-6bb7b6db47-srl48\" (UID: \"0deb5e73-3948-40cb-8dca-c1ebcf8c117f\") " pod="openshift-controller-manager/controller-manager-6bb7b6db47-srl48" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.689106 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/0deb5e73-3948-40cb-8dca-c1ebcf8c117f-client-ca\") pod \"controller-manager-6bb7b6db47-srl48\" (UID: \"0deb5e73-3948-40cb-8dca-c1ebcf8c117f\") " pod="openshift-controller-manager/controller-manager-6bb7b6db47-srl48" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.689417 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/0deb5e73-3948-40cb-8dca-c1ebcf8c117f-proxy-ca-bundles\") pod \"controller-manager-6bb7b6db47-srl48\" (UID: \"0deb5e73-3948-40cb-8dca-c1ebcf8c117f\") " pod="openshift-controller-manager/controller-manager-6bb7b6db47-srl48" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.689589 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/0deb5e73-3948-40cb-8dca-c1ebcf8c117f-config\") pod \"controller-manager-6bb7b6db47-srl48\" (UID: \"0deb5e73-3948-40cb-8dca-c1ebcf8c117f\") " pod="openshift-controller-manager/controller-manager-6bb7b6db47-srl48" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.694779 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0deb5e73-3948-40cb-8dca-c1ebcf8c117f-serving-cert\") pod \"controller-manager-6bb7b6db47-srl48\" (UID: \"0deb5e73-3948-40cb-8dca-c1ebcf8c117f\") " pod="openshift-controller-manager/controller-manager-6bb7b6db47-srl48" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.705482 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f955r\" (UniqueName: \"kubernetes.io/projected/0deb5e73-3948-40cb-8dca-c1ebcf8c117f-kube-api-access-f955r\") pod \"controller-manager-6bb7b6db47-srl48\" (UID: \"0deb5e73-3948-40cb-8dca-c1ebcf8c117f\") " pod="openshift-controller-manager/controller-manager-6bb7b6db47-srl48" Dec 09 15:36:57 crc kubenswrapper[4894]: I1209 15:36:57.775689 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-6bb7b6db47-srl48" Dec 09 15:36:58 crc kubenswrapper[4894]: I1209 15:36:58.106110 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Dec 09 15:36:58 crc kubenswrapper[4894]: I1209 15:36:58.154915 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-6bb7b6db47-srl48"] Dec 09 15:36:58 crc kubenswrapper[4894]: W1209 15:36:58.162967 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0deb5e73_3948_40cb_8dca_c1ebcf8c117f.slice/crio-f1aa56c94713807cd71e6e7d0413cad64fddecb2c0c365b2518d86467b48875f WatchSource:0}: Error finding container f1aa56c94713807cd71e6e7d0413cad64fddecb2c0c365b2518d86467b48875f: Status 404 returned error can't find the container with id f1aa56c94713807cd71e6e7d0413cad64fddecb2c0c365b2518d86467b48875f Dec 09 15:36:58 crc kubenswrapper[4894]: I1209 15:36:58.546462 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bb7b6db47-srl48" event={"ID":"0deb5e73-3948-40cb-8dca-c1ebcf8c117f","Type":"ContainerStarted","Data":"50d314cb7b8305eb16a57ce513be72ea904009515d11496a78173ab615f4ecf4"} Dec 09 15:36:58 crc kubenswrapper[4894]: I1209 15:36:58.546518 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-6bb7b6db47-srl48" event={"ID":"0deb5e73-3948-40cb-8dca-c1ebcf8c117f","Type":"ContainerStarted","Data":"f1aa56c94713807cd71e6e7d0413cad64fddecb2c0c365b2518d86467b48875f"} Dec 09 15:36:58 crc kubenswrapper[4894]: I1209 15:36:58.580248 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-6bb7b6db47-srl48" podStartSLOduration=6.580226362 podStartE2EDuration="6.580226362s" podCreationTimestamp="2025-12-09 15:36:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:36:58.575648969 +0000 UTC m=+312.894859648" watchObservedRunningTime="2025-12-09 15:36:58.580226362 +0000 UTC m=+312.899437031" Dec 09 15:36:59 crc kubenswrapper[4894]: I1209 15:36:59.551055 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-6bb7b6db47-srl48" Dec 09 15:36:59 crc kubenswrapper[4894]: I1209 15:36:59.555732 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-6bb7b6db47-srl48" Dec 09 15:37:12 crc kubenswrapper[4894]: I1209 15:37:12.257674 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:37:12 crc kubenswrapper[4894]: I1209 15:37:12.258209 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.568951 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tjk2z"] Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.570373 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.579394 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tjk2z"] Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.693656 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/16de5fc8-8ef2-443c-b6eb-fc262fcbb280-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.693707 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/16de5fc8-8ef2-443c-b6eb-fc262fcbb280-registry-tls\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.693740 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/16de5fc8-8ef2-443c-b6eb-fc262fcbb280-bound-sa-token\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.693768 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/16de5fc8-8ef2-443c-b6eb-fc262fcbb280-registry-certificates\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.693807 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/16de5fc8-8ef2-443c-b6eb-fc262fcbb280-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.693832 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16de5fc8-8ef2-443c-b6eb-fc262fcbb280-trusted-ca\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.693957 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdksq\" (UniqueName: \"kubernetes.io/projected/16de5fc8-8ef2-443c-b6eb-fc262fcbb280-kube-api-access-rdksq\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.694029 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.722001 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.795410 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/16de5fc8-8ef2-443c-b6eb-fc262fcbb280-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.795475 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/16de5fc8-8ef2-443c-b6eb-fc262fcbb280-registry-tls\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.795921 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/16de5fc8-8ef2-443c-b6eb-fc262fcbb280-registry-certificates\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.795961 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/16de5fc8-8ef2-443c-b6eb-fc262fcbb280-bound-sa-token\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.796008 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/16de5fc8-8ef2-443c-b6eb-fc262fcbb280-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.796046 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16de5fc8-8ef2-443c-b6eb-fc262fcbb280-trusted-ca\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.796075 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdksq\" (UniqueName: \"kubernetes.io/projected/16de5fc8-8ef2-443c-b6eb-fc262fcbb280-kube-api-access-rdksq\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.796121 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/16de5fc8-8ef2-443c-b6eb-fc262fcbb280-ca-trust-extracted\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.797987 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/16de5fc8-8ef2-443c-b6eb-fc262fcbb280-trusted-ca\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.798128 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/16de5fc8-8ef2-443c-b6eb-fc262fcbb280-registry-certificates\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.804429 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/16de5fc8-8ef2-443c-b6eb-fc262fcbb280-registry-tls\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.804496 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/16de5fc8-8ef2-443c-b6eb-fc262fcbb280-installation-pull-secrets\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.812953 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/16de5fc8-8ef2-443c-b6eb-fc262fcbb280-bound-sa-token\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.815933 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdksq\" (UniqueName: \"kubernetes.io/projected/16de5fc8-8ef2-443c-b6eb-fc262fcbb280-kube-api-access-rdksq\") pod \"image-registry-66df7c8f76-tjk2z\" (UID: \"16de5fc8-8ef2-443c-b6eb-fc262fcbb280\") " pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:22 crc kubenswrapper[4894]: I1209 15:37:22.897225 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.398267 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8hxbx"] Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.399692 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8hxbx" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.403513 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.418665 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8hxbx"] Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.482602 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-tjk2z"] Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.505430 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83805fb6-3997-4aba-a8ca-15c24e4afe4c-utilities\") pod \"redhat-operators-8hxbx\" (UID: \"83805fb6-3997-4aba-a8ca-15c24e4afe4c\") " pod="openshift-marketplace/redhat-operators-8hxbx" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.505493 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83805fb6-3997-4aba-a8ca-15c24e4afe4c-catalog-content\") pod \"redhat-operators-8hxbx\" (UID: \"83805fb6-3997-4aba-a8ca-15c24e4afe4c\") " pod="openshift-marketplace/redhat-operators-8hxbx" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.505630 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mgwhb\" (UniqueName: \"kubernetes.io/projected/83805fb6-3997-4aba-a8ca-15c24e4afe4c-kube-api-access-mgwhb\") pod \"redhat-operators-8hxbx\" (UID: \"83805fb6-3997-4aba-a8ca-15c24e4afe4c\") " pod="openshift-marketplace/redhat-operators-8hxbx" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.599625 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zg6pn"] Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.600860 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zg6pn" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.603592 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.607418 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83805fb6-3997-4aba-a8ca-15c24e4afe4c-utilities\") pod \"redhat-operators-8hxbx\" (UID: \"83805fb6-3997-4aba-a8ca-15c24e4afe4c\") " pod="openshift-marketplace/redhat-operators-8hxbx" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.607486 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83805fb6-3997-4aba-a8ca-15c24e4afe4c-catalog-content\") pod \"redhat-operators-8hxbx\" (UID: \"83805fb6-3997-4aba-a8ca-15c24e4afe4c\") " pod="openshift-marketplace/redhat-operators-8hxbx" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.607570 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mgwhb\" (UniqueName: \"kubernetes.io/projected/83805fb6-3997-4aba-a8ca-15c24e4afe4c-kube-api-access-mgwhb\") pod \"redhat-operators-8hxbx\" (UID: \"83805fb6-3997-4aba-a8ca-15c24e4afe4c\") " pod="openshift-marketplace/redhat-operators-8hxbx" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.607938 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83805fb6-3997-4aba-a8ca-15c24e4afe4c-utilities\") pod \"redhat-operators-8hxbx\" (UID: \"83805fb6-3997-4aba-a8ca-15c24e4afe4c\") " pod="openshift-marketplace/redhat-operators-8hxbx" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.608305 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83805fb6-3997-4aba-a8ca-15c24e4afe4c-catalog-content\") pod \"redhat-operators-8hxbx\" (UID: \"83805fb6-3997-4aba-a8ca-15c24e4afe4c\") " pod="openshift-marketplace/redhat-operators-8hxbx" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.608568 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zg6pn"] Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.630289 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mgwhb\" (UniqueName: \"kubernetes.io/projected/83805fb6-3997-4aba-a8ca-15c24e4afe4c-kube-api-access-mgwhb\") pod \"redhat-operators-8hxbx\" (UID: \"83805fb6-3997-4aba-a8ca-15c24e4afe4c\") " pod="openshift-marketplace/redhat-operators-8hxbx" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.689972 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" event={"ID":"16de5fc8-8ef2-443c-b6eb-fc262fcbb280","Type":"ContainerStarted","Data":"6a9477d167cbd44b7fd11b3955d3351b5f80a57908a6b3f5fb92cb5208fc929c"} Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.690018 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" event={"ID":"16de5fc8-8ef2-443c-b6eb-fc262fcbb280","Type":"ContainerStarted","Data":"31cda92162ddc081b5a56c24b26128ae9a9c2cb2aefad2591eacdfc28ebc199e"} Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.690126 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.709769 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" podStartSLOduration=1.7097517500000001 podStartE2EDuration="1.70975175s" podCreationTimestamp="2025-12-09 15:37:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:37:23.708331876 +0000 UTC m=+338.027542545" watchObservedRunningTime="2025-12-09 15:37:23.70975175 +0000 UTC m=+338.028962419" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.714790 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/992e28b8-b2cb-4795-b27e-76b87af880c3-catalog-content\") pod \"redhat-marketplace-zg6pn\" (UID: \"992e28b8-b2cb-4795-b27e-76b87af880c3\") " pod="openshift-marketplace/redhat-marketplace-zg6pn" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.714877 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s57nf\" (UniqueName: \"kubernetes.io/projected/992e28b8-b2cb-4795-b27e-76b87af880c3-kube-api-access-s57nf\") pod \"redhat-marketplace-zg6pn\" (UID: \"992e28b8-b2cb-4795-b27e-76b87af880c3\") " pod="openshift-marketplace/redhat-marketplace-zg6pn" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.714906 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/992e28b8-b2cb-4795-b27e-76b87af880c3-utilities\") pod \"redhat-marketplace-zg6pn\" (UID: \"992e28b8-b2cb-4795-b27e-76b87af880c3\") " pod="openshift-marketplace/redhat-marketplace-zg6pn" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.725539 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8hxbx" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.816980 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s57nf\" (UniqueName: \"kubernetes.io/projected/992e28b8-b2cb-4795-b27e-76b87af880c3-kube-api-access-s57nf\") pod \"redhat-marketplace-zg6pn\" (UID: \"992e28b8-b2cb-4795-b27e-76b87af880c3\") " pod="openshift-marketplace/redhat-marketplace-zg6pn" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.817337 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/992e28b8-b2cb-4795-b27e-76b87af880c3-utilities\") pod \"redhat-marketplace-zg6pn\" (UID: \"992e28b8-b2cb-4795-b27e-76b87af880c3\") " pod="openshift-marketplace/redhat-marketplace-zg6pn" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.817474 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/992e28b8-b2cb-4795-b27e-76b87af880c3-catalog-content\") pod \"redhat-marketplace-zg6pn\" (UID: \"992e28b8-b2cb-4795-b27e-76b87af880c3\") " pod="openshift-marketplace/redhat-marketplace-zg6pn" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.817934 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/992e28b8-b2cb-4795-b27e-76b87af880c3-catalog-content\") pod \"redhat-marketplace-zg6pn\" (UID: \"992e28b8-b2cb-4795-b27e-76b87af880c3\") " pod="openshift-marketplace/redhat-marketplace-zg6pn" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.818114 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/992e28b8-b2cb-4795-b27e-76b87af880c3-utilities\") pod \"redhat-marketplace-zg6pn\" (UID: \"992e28b8-b2cb-4795-b27e-76b87af880c3\") " pod="openshift-marketplace/redhat-marketplace-zg6pn" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.849400 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s57nf\" (UniqueName: \"kubernetes.io/projected/992e28b8-b2cb-4795-b27e-76b87af880c3-kube-api-access-s57nf\") pod \"redhat-marketplace-zg6pn\" (UID: \"992e28b8-b2cb-4795-b27e-76b87af880c3\") " pod="openshift-marketplace/redhat-marketplace-zg6pn" Dec 09 15:37:23 crc kubenswrapper[4894]: I1209 15:37:23.924296 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zg6pn" Dec 09 15:37:24 crc kubenswrapper[4894]: I1209 15:37:24.193635 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8hxbx"] Dec 09 15:37:24 crc kubenswrapper[4894]: W1209 15:37:24.196879 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod83805fb6_3997_4aba_a8ca_15c24e4afe4c.slice/crio-5b12050d435e6d73449e1352c9a5253fd3b12544e1678d44e241649aaa2356da WatchSource:0}: Error finding container 5b12050d435e6d73449e1352c9a5253fd3b12544e1678d44e241649aaa2356da: Status 404 returned error can't find the container with id 5b12050d435e6d73449e1352c9a5253fd3b12544e1678d44e241649aaa2356da Dec 09 15:37:24 crc kubenswrapper[4894]: I1209 15:37:24.493507 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zg6pn"] Dec 09 15:37:24 crc kubenswrapper[4894]: W1209 15:37:24.500292 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod992e28b8_b2cb_4795_b27e_76b87af880c3.slice/crio-d9350dbfb1bd78a49d7a91439c46de416c1698faf76daa4ff0eb6f70d51514a0 WatchSource:0}: Error finding container d9350dbfb1bd78a49d7a91439c46de416c1698faf76daa4ff0eb6f70d51514a0: Status 404 returned error can't find the container with id d9350dbfb1bd78a49d7a91439c46de416c1698faf76daa4ff0eb6f70d51514a0 Dec 09 15:37:24 crc kubenswrapper[4894]: I1209 15:37:24.697556 4894 generic.go:334] "Generic (PLEG): container finished" podID="83805fb6-3997-4aba-a8ca-15c24e4afe4c" containerID="17050e0f7ad4a2484af86173376dc1059f8c6ec77c689d9b8411f81670f51d57" exitCode=0 Dec 09 15:37:24 crc kubenswrapper[4894]: I1209 15:37:24.698709 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8hxbx" event={"ID":"83805fb6-3997-4aba-a8ca-15c24e4afe4c","Type":"ContainerDied","Data":"17050e0f7ad4a2484af86173376dc1059f8c6ec77c689d9b8411f81670f51d57"} Dec 09 15:37:24 crc kubenswrapper[4894]: I1209 15:37:24.698763 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8hxbx" event={"ID":"83805fb6-3997-4aba-a8ca-15c24e4afe4c","Type":"ContainerStarted","Data":"5b12050d435e6d73449e1352c9a5253fd3b12544e1678d44e241649aaa2356da"} Dec 09 15:37:24 crc kubenswrapper[4894]: I1209 15:37:24.702653 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zg6pn" event={"ID":"992e28b8-b2cb-4795-b27e-76b87af880c3","Type":"ContainerStarted","Data":"d9350dbfb1bd78a49d7a91439c46de416c1698faf76daa4ff0eb6f70d51514a0"} Dec 09 15:37:25 crc kubenswrapper[4894]: I1209 15:37:25.709356 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8hxbx" event={"ID":"83805fb6-3997-4aba-a8ca-15c24e4afe4c","Type":"ContainerStarted","Data":"7729c3bbe467731935f01097c7facf6bc0e2557877d7ba63489131930e381403"} Dec 09 15:37:25 crc kubenswrapper[4894]: I1209 15:37:25.712982 4894 generic.go:334] "Generic (PLEG): container finished" podID="992e28b8-b2cb-4795-b27e-76b87af880c3" containerID="0795f598e6557c0341b2fb4ae3f1c355c3026c965cd37a521ed816b0c0e8ffc2" exitCode=0 Dec 09 15:37:25 crc kubenswrapper[4894]: I1209 15:37:25.713023 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zg6pn" event={"ID":"992e28b8-b2cb-4795-b27e-76b87af880c3","Type":"ContainerDied","Data":"0795f598e6557c0341b2fb4ae3f1c355c3026c965cd37a521ed816b0c0e8ffc2"} Dec 09 15:37:25 crc kubenswrapper[4894]: I1209 15:37:25.800931 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-9dcgf"] Dec 09 15:37:25 crc kubenswrapper[4894]: I1209 15:37:25.802341 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9dcgf" Dec 09 15:37:25 crc kubenswrapper[4894]: I1209 15:37:25.804615 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Dec 09 15:37:25 crc kubenswrapper[4894]: I1209 15:37:25.832124 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9dcgf"] Dec 09 15:37:25 crc kubenswrapper[4894]: I1209 15:37:25.942206 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfcbb982-3316-4cda-b012-0813b7fd4259-catalog-content\") pod \"certified-operators-9dcgf\" (UID: \"bfcbb982-3316-4cda-b012-0813b7fd4259\") " pod="openshift-marketplace/certified-operators-9dcgf" Dec 09 15:37:25 crc kubenswrapper[4894]: I1209 15:37:25.942328 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfcbb982-3316-4cda-b012-0813b7fd4259-utilities\") pod \"certified-operators-9dcgf\" (UID: \"bfcbb982-3316-4cda-b012-0813b7fd4259\") " pod="openshift-marketplace/certified-operators-9dcgf" Dec 09 15:37:25 crc kubenswrapper[4894]: I1209 15:37:25.942366 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mncp4\" (UniqueName: \"kubernetes.io/projected/bfcbb982-3316-4cda-b012-0813b7fd4259-kube-api-access-mncp4\") pod \"certified-operators-9dcgf\" (UID: \"bfcbb982-3316-4cda-b012-0813b7fd4259\") " pod="openshift-marketplace/certified-operators-9dcgf" Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.006454 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-rb4sl"] Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.008830 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rb4sl" Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.012927 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.013162 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rb4sl"] Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.043894 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfcbb982-3316-4cda-b012-0813b7fd4259-utilities\") pod \"certified-operators-9dcgf\" (UID: \"bfcbb982-3316-4cda-b012-0813b7fd4259\") " pod="openshift-marketplace/certified-operators-9dcgf" Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.043954 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mncp4\" (UniqueName: \"kubernetes.io/projected/bfcbb982-3316-4cda-b012-0813b7fd4259-kube-api-access-mncp4\") pod \"certified-operators-9dcgf\" (UID: \"bfcbb982-3316-4cda-b012-0813b7fd4259\") " pod="openshift-marketplace/certified-operators-9dcgf" Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.043994 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfcbb982-3316-4cda-b012-0813b7fd4259-catalog-content\") pod \"certified-operators-9dcgf\" (UID: \"bfcbb982-3316-4cda-b012-0813b7fd4259\") " pod="openshift-marketplace/certified-operators-9dcgf" Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.044467 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfcbb982-3316-4cda-b012-0813b7fd4259-catalog-content\") pod \"certified-operators-9dcgf\" (UID: \"bfcbb982-3316-4cda-b012-0813b7fd4259\") " pod="openshift-marketplace/certified-operators-9dcgf" Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.044735 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfcbb982-3316-4cda-b012-0813b7fd4259-utilities\") pod \"certified-operators-9dcgf\" (UID: \"bfcbb982-3316-4cda-b012-0813b7fd4259\") " pod="openshift-marketplace/certified-operators-9dcgf" Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.067613 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mncp4\" (UniqueName: \"kubernetes.io/projected/bfcbb982-3316-4cda-b012-0813b7fd4259-kube-api-access-mncp4\") pod \"certified-operators-9dcgf\" (UID: \"bfcbb982-3316-4cda-b012-0813b7fd4259\") " pod="openshift-marketplace/certified-operators-9dcgf" Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.128510 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9dcgf" Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.145442 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b469772b-9a9b-43dd-92ea-059528977c3a-catalog-content\") pod \"community-operators-rb4sl\" (UID: \"b469772b-9a9b-43dd-92ea-059528977c3a\") " pod="openshift-marketplace/community-operators-rb4sl" Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.145503 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b469772b-9a9b-43dd-92ea-059528977c3a-utilities\") pod \"community-operators-rb4sl\" (UID: \"b469772b-9a9b-43dd-92ea-059528977c3a\") " pod="openshift-marketplace/community-operators-rb4sl" Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.145607 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tt8wr\" (UniqueName: \"kubernetes.io/projected/b469772b-9a9b-43dd-92ea-059528977c3a-kube-api-access-tt8wr\") pod \"community-operators-rb4sl\" (UID: \"b469772b-9a9b-43dd-92ea-059528977c3a\") " pod="openshift-marketplace/community-operators-rb4sl" Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.246579 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b469772b-9a9b-43dd-92ea-059528977c3a-catalog-content\") pod \"community-operators-rb4sl\" (UID: \"b469772b-9a9b-43dd-92ea-059528977c3a\") " pod="openshift-marketplace/community-operators-rb4sl" Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.247097 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b469772b-9a9b-43dd-92ea-059528977c3a-utilities\") pod \"community-operators-rb4sl\" (UID: \"b469772b-9a9b-43dd-92ea-059528977c3a\") " pod="openshift-marketplace/community-operators-rb4sl" Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.247146 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tt8wr\" (UniqueName: \"kubernetes.io/projected/b469772b-9a9b-43dd-92ea-059528977c3a-kube-api-access-tt8wr\") pod \"community-operators-rb4sl\" (UID: \"b469772b-9a9b-43dd-92ea-059528977c3a\") " pod="openshift-marketplace/community-operators-rb4sl" Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.247560 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b469772b-9a9b-43dd-92ea-059528977c3a-catalog-content\") pod \"community-operators-rb4sl\" (UID: \"b469772b-9a9b-43dd-92ea-059528977c3a\") " pod="openshift-marketplace/community-operators-rb4sl" Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.247851 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b469772b-9a9b-43dd-92ea-059528977c3a-utilities\") pod \"community-operators-rb4sl\" (UID: \"b469772b-9a9b-43dd-92ea-059528977c3a\") " pod="openshift-marketplace/community-operators-rb4sl" Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.267275 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tt8wr\" (UniqueName: \"kubernetes.io/projected/b469772b-9a9b-43dd-92ea-059528977c3a-kube-api-access-tt8wr\") pod \"community-operators-rb4sl\" (UID: \"b469772b-9a9b-43dd-92ea-059528977c3a\") " pod="openshift-marketplace/community-operators-rb4sl" Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.386413 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-rb4sl" Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.533352 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-9dcgf"] Dec 09 15:37:26 crc kubenswrapper[4894]: W1209 15:37:26.540682 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbfcbb982_3316_4cda_b012_0813b7fd4259.slice/crio-64cde353590a62be16c6ac0542938adc7bdb6b02c8a628def25c07023579ae7c WatchSource:0}: Error finding container 64cde353590a62be16c6ac0542938adc7bdb6b02c8a628def25c07023579ae7c: Status 404 returned error can't find the container with id 64cde353590a62be16c6ac0542938adc7bdb6b02c8a628def25c07023579ae7c Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.718279 4894 generic.go:334] "Generic (PLEG): container finished" podID="bfcbb982-3316-4cda-b012-0813b7fd4259" containerID="a95235df854dd9332191b34d0879a6e97824ae9b9a73e4ccd97151ce443be31b" exitCode=0 Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.718549 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9dcgf" event={"ID":"bfcbb982-3316-4cda-b012-0813b7fd4259","Type":"ContainerDied","Data":"a95235df854dd9332191b34d0879a6e97824ae9b9a73e4ccd97151ce443be31b"} Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.718682 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9dcgf" event={"ID":"bfcbb982-3316-4cda-b012-0813b7fd4259","Type":"ContainerStarted","Data":"64cde353590a62be16c6ac0542938adc7bdb6b02c8a628def25c07023579ae7c"} Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.722865 4894 generic.go:334] "Generic (PLEG): container finished" podID="992e28b8-b2cb-4795-b27e-76b87af880c3" containerID="cbb362f09d403ea4dbfabfd98f832757396ddd152ec3084984ec9745e1874525" exitCode=0 Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.722943 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zg6pn" event={"ID":"992e28b8-b2cb-4795-b27e-76b87af880c3","Type":"ContainerDied","Data":"cbb362f09d403ea4dbfabfd98f832757396ddd152ec3084984ec9745e1874525"} Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.725670 4894 generic.go:334] "Generic (PLEG): container finished" podID="83805fb6-3997-4aba-a8ca-15c24e4afe4c" containerID="7729c3bbe467731935f01097c7facf6bc0e2557877d7ba63489131930e381403" exitCode=0 Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.725744 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8hxbx" event={"ID":"83805fb6-3997-4aba-a8ca-15c24e4afe4c","Type":"ContainerDied","Data":"7729c3bbe467731935f01097c7facf6bc0e2557877d7ba63489131930e381403"} Dec 09 15:37:26 crc kubenswrapper[4894]: I1209 15:37:26.784941 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-rb4sl"] Dec 09 15:37:27 crc kubenswrapper[4894]: I1209 15:37:27.734685 4894 generic.go:334] "Generic (PLEG): container finished" podID="bfcbb982-3316-4cda-b012-0813b7fd4259" containerID="89ddccc4bb8b2ff1669dcd7f11f8afd573a55d4a68090eb0af54957223678427" exitCode=0 Dec 09 15:37:27 crc kubenswrapper[4894]: I1209 15:37:27.734778 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9dcgf" event={"ID":"bfcbb982-3316-4cda-b012-0813b7fd4259","Type":"ContainerDied","Data":"89ddccc4bb8b2ff1669dcd7f11f8afd573a55d4a68090eb0af54957223678427"} Dec 09 15:37:27 crc kubenswrapper[4894]: I1209 15:37:27.737986 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zg6pn" event={"ID":"992e28b8-b2cb-4795-b27e-76b87af880c3","Type":"ContainerStarted","Data":"3f0b42b9ca5e3e43a6dec5239ac2d97b669c194436472c534abae152ea54c22d"} Dec 09 15:37:27 crc kubenswrapper[4894]: I1209 15:37:27.741381 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8hxbx" event={"ID":"83805fb6-3997-4aba-a8ca-15c24e4afe4c","Type":"ContainerStarted","Data":"f896b45b40fb7bc155a8021c33fa428209e6e4658fa88b152ff574a1141d6e9a"} Dec 09 15:37:27 crc kubenswrapper[4894]: I1209 15:37:27.743132 4894 generic.go:334] "Generic (PLEG): container finished" podID="b469772b-9a9b-43dd-92ea-059528977c3a" containerID="d83c4c2de6dbf40026ab84d900453c1abc7e536a16130ed3a4a02a2569a6d421" exitCode=0 Dec 09 15:37:27 crc kubenswrapper[4894]: I1209 15:37:27.743175 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rb4sl" event={"ID":"b469772b-9a9b-43dd-92ea-059528977c3a","Type":"ContainerDied","Data":"d83c4c2de6dbf40026ab84d900453c1abc7e536a16130ed3a4a02a2569a6d421"} Dec 09 15:37:27 crc kubenswrapper[4894]: I1209 15:37:27.743198 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rb4sl" event={"ID":"b469772b-9a9b-43dd-92ea-059528977c3a","Type":"ContainerStarted","Data":"5a447ba8d96292646db69814f849e5176a1893e95cb365e4c206fd5410cc90a7"} Dec 09 15:37:27 crc kubenswrapper[4894]: I1209 15:37:27.782198 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zg6pn" podStartSLOduration=3.285636937 podStartE2EDuration="4.78217727s" podCreationTimestamp="2025-12-09 15:37:23 +0000 UTC" firstStartedPulling="2025-12-09 15:37:25.714263491 +0000 UTC m=+340.033474160" lastFinishedPulling="2025-12-09 15:37:27.210803824 +0000 UTC m=+341.530014493" observedRunningTime="2025-12-09 15:37:27.780822788 +0000 UTC m=+342.100033487" watchObservedRunningTime="2025-12-09 15:37:27.78217727 +0000 UTC m=+342.101387939" Dec 09 15:37:27 crc kubenswrapper[4894]: I1209 15:37:27.798626 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8hxbx" podStartSLOduration=2.156569697 podStartE2EDuration="4.798608317s" podCreationTimestamp="2025-12-09 15:37:23 +0000 UTC" firstStartedPulling="2025-12-09 15:37:24.699706286 +0000 UTC m=+339.018916955" lastFinishedPulling="2025-12-09 15:37:27.341744886 +0000 UTC m=+341.660955575" observedRunningTime="2025-12-09 15:37:27.796499893 +0000 UTC m=+342.115710552" watchObservedRunningTime="2025-12-09 15:37:27.798608317 +0000 UTC m=+342.117818986" Dec 09 15:37:28 crc kubenswrapper[4894]: I1209 15:37:28.749023 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9dcgf" event={"ID":"bfcbb982-3316-4cda-b012-0813b7fd4259","Type":"ContainerStarted","Data":"1d7593f67faeea7eaf084f0f7c34171f41fdcb92096a5b20b7de1312d1d4c7e6"} Dec 09 15:37:28 crc kubenswrapper[4894]: I1209 15:37:28.768918 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-9dcgf" podStartSLOduration=2.2232469679999998 podStartE2EDuration="3.768902107s" podCreationTimestamp="2025-12-09 15:37:25 +0000 UTC" firstStartedPulling="2025-12-09 15:37:26.719897472 +0000 UTC m=+341.039108141" lastFinishedPulling="2025-12-09 15:37:28.265552611 +0000 UTC m=+342.584763280" observedRunningTime="2025-12-09 15:37:28.767630507 +0000 UTC m=+343.086841186" watchObservedRunningTime="2025-12-09 15:37:28.768902107 +0000 UTC m=+343.088112776" Dec 09 15:37:29 crc kubenswrapper[4894]: I1209 15:37:29.756589 4894 generic.go:334] "Generic (PLEG): container finished" podID="b469772b-9a9b-43dd-92ea-059528977c3a" containerID="1d9b62d16133a2e83ab89a81d892f69723d10984e0b787bf9a95fb3568a3ec8c" exitCode=0 Dec 09 15:37:29 crc kubenswrapper[4894]: I1209 15:37:29.756772 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rb4sl" event={"ID":"b469772b-9a9b-43dd-92ea-059528977c3a","Type":"ContainerDied","Data":"1d9b62d16133a2e83ab89a81d892f69723d10984e0b787bf9a95fb3568a3ec8c"} Dec 09 15:37:30 crc kubenswrapper[4894]: I1209 15:37:30.763698 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-rb4sl" event={"ID":"b469772b-9a9b-43dd-92ea-059528977c3a","Type":"ContainerStarted","Data":"d6c3a09b3073535b9be6fb93eeef75e804464d942a5353f695a17db137bddcc2"} Dec 09 15:37:30 crc kubenswrapper[4894]: I1209 15:37:30.784354 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-rb4sl" podStartSLOduration=3.3164039 podStartE2EDuration="5.784324305s" podCreationTimestamp="2025-12-09 15:37:25 +0000 UTC" firstStartedPulling="2025-12-09 15:37:27.744676793 +0000 UTC m=+342.063887462" lastFinishedPulling="2025-12-09 15:37:30.212597198 +0000 UTC m=+344.531807867" observedRunningTime="2025-12-09 15:37:30.781227849 +0000 UTC m=+345.100438528" watchObservedRunningTime="2025-12-09 15:37:30.784324305 +0000 UTC m=+345.103534974" Dec 09 15:37:33 crc kubenswrapper[4894]: I1209 15:37:33.447689 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-g79bw"] Dec 09 15:37:33 crc kubenswrapper[4894]: I1209 15:37:33.725813 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8hxbx" Dec 09 15:37:33 crc kubenswrapper[4894]: I1209 15:37:33.725889 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8hxbx" Dec 09 15:37:33 crc kubenswrapper[4894]: I1209 15:37:33.783987 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8hxbx" Dec 09 15:37:33 crc kubenswrapper[4894]: I1209 15:37:33.852405 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8hxbx" Dec 09 15:37:33 crc kubenswrapper[4894]: I1209 15:37:33.924598 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zg6pn" Dec 09 15:37:33 crc kubenswrapper[4894]: I1209 15:37:33.924666 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zg6pn" Dec 09 15:37:33 crc kubenswrapper[4894]: I1209 15:37:33.960018 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zg6pn" Dec 09 15:37:34 crc kubenswrapper[4894]: I1209 15:37:34.833510 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zg6pn" Dec 09 15:37:36 crc kubenswrapper[4894]: I1209 15:37:36.129573 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-9dcgf" Dec 09 15:37:36 crc kubenswrapper[4894]: I1209 15:37:36.129839 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-9dcgf" Dec 09 15:37:36 crc kubenswrapper[4894]: I1209 15:37:36.183964 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-9dcgf" Dec 09 15:37:36 crc kubenswrapper[4894]: I1209 15:37:36.394897 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-rb4sl" Dec 09 15:37:36 crc kubenswrapper[4894]: I1209 15:37:36.394956 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-rb4sl" Dec 09 15:37:36 crc kubenswrapper[4894]: I1209 15:37:36.454429 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-rb4sl" Dec 09 15:37:36 crc kubenswrapper[4894]: I1209 15:37:36.836558 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-9dcgf" Dec 09 15:37:36 crc kubenswrapper[4894]: I1209 15:37:36.850013 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-rb4sl" Dec 09 15:37:42 crc kubenswrapper[4894]: I1209 15:37:42.258082 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:37:42 crc kubenswrapper[4894]: I1209 15:37:42.258422 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:37:42 crc kubenswrapper[4894]: I1209 15:37:42.902881 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-tjk2z" Dec 09 15:37:42 crc kubenswrapper[4894]: I1209 15:37:42.946051 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k8jmv"] Dec 09 15:37:47 crc kubenswrapper[4894]: I1209 15:37:47.023145 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd"] Dec 09 15:37:47 crc kubenswrapper[4894]: I1209 15:37:47.023851 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" podUID="209ad566-6ec9-44ba-9a0a-bc62cbfea583" containerName="route-controller-manager" containerID="cri-o://6a5a84367ea89e1aa00cd610ed60fa492af84bf3b3be6523d156055c59749b1a" gracePeriod=30 Dec 09 15:37:49 crc kubenswrapper[4894]: I1209 15:37:49.871410 4894 generic.go:334] "Generic (PLEG): container finished" podID="209ad566-6ec9-44ba-9a0a-bc62cbfea583" containerID="6a5a84367ea89e1aa00cd610ed60fa492af84bf3b3be6523d156055c59749b1a" exitCode=0 Dec 09 15:37:49 crc kubenswrapper[4894]: I1209 15:37:49.871594 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" event={"ID":"209ad566-6ec9-44ba-9a0a-bc62cbfea583","Type":"ContainerDied","Data":"6a5a84367ea89e1aa00cd610ed60fa492af84bf3b3be6523d156055c59749b1a"} Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.066272 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.097533 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c66dcf44f-f6gkh"] Dec 09 15:37:50 crc kubenswrapper[4894]: E1209 15:37:50.098373 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="209ad566-6ec9-44ba-9a0a-bc62cbfea583" containerName="route-controller-manager" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.098397 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="209ad566-6ec9-44ba-9a0a-bc62cbfea583" containerName="route-controller-manager" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.098511 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="209ad566-6ec9-44ba-9a0a-bc62cbfea583" containerName="route-controller-manager" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.099080 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c66dcf44f-f6gkh" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.120123 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c66dcf44f-f6gkh"] Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.238295 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/209ad566-6ec9-44ba-9a0a-bc62cbfea583-serving-cert\") pod \"209ad566-6ec9-44ba-9a0a-bc62cbfea583\" (UID: \"209ad566-6ec9-44ba-9a0a-bc62cbfea583\") " Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.238363 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/209ad566-6ec9-44ba-9a0a-bc62cbfea583-client-ca\") pod \"209ad566-6ec9-44ba-9a0a-bc62cbfea583\" (UID: \"209ad566-6ec9-44ba-9a0a-bc62cbfea583\") " Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.238395 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/209ad566-6ec9-44ba-9a0a-bc62cbfea583-config\") pod \"209ad566-6ec9-44ba-9a0a-bc62cbfea583\" (UID: \"209ad566-6ec9-44ba-9a0a-bc62cbfea583\") " Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.238428 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qf67l\" (UniqueName: \"kubernetes.io/projected/209ad566-6ec9-44ba-9a0a-bc62cbfea583-kube-api-access-qf67l\") pod \"209ad566-6ec9-44ba-9a0a-bc62cbfea583\" (UID: \"209ad566-6ec9-44ba-9a0a-bc62cbfea583\") " Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.238603 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af8e7802-83b6-41c9-936c-b4837ecdd732-client-ca\") pod \"route-controller-manager-c66dcf44f-f6gkh\" (UID: \"af8e7802-83b6-41c9-936c-b4837ecdd732\") " pod="openshift-route-controller-manager/route-controller-manager-c66dcf44f-f6gkh" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.238630 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af8e7802-83b6-41c9-936c-b4837ecdd732-serving-cert\") pod \"route-controller-manager-c66dcf44f-f6gkh\" (UID: \"af8e7802-83b6-41c9-936c-b4837ecdd732\") " pod="openshift-route-controller-manager/route-controller-manager-c66dcf44f-f6gkh" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.238692 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af8e7802-83b6-41c9-936c-b4837ecdd732-config\") pod \"route-controller-manager-c66dcf44f-f6gkh\" (UID: \"af8e7802-83b6-41c9-936c-b4837ecdd732\") " pod="openshift-route-controller-manager/route-controller-manager-c66dcf44f-f6gkh" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.238708 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v4ww8\" (UniqueName: \"kubernetes.io/projected/af8e7802-83b6-41c9-936c-b4837ecdd732-kube-api-access-v4ww8\") pod \"route-controller-manager-c66dcf44f-f6gkh\" (UID: \"af8e7802-83b6-41c9-936c-b4837ecdd732\") " pod="openshift-route-controller-manager/route-controller-manager-c66dcf44f-f6gkh" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.239187 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/209ad566-6ec9-44ba-9a0a-bc62cbfea583-client-ca" (OuterVolumeSpecName: "client-ca") pod "209ad566-6ec9-44ba-9a0a-bc62cbfea583" (UID: "209ad566-6ec9-44ba-9a0a-bc62cbfea583"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.239387 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/209ad566-6ec9-44ba-9a0a-bc62cbfea583-config" (OuterVolumeSpecName: "config") pod "209ad566-6ec9-44ba-9a0a-bc62cbfea583" (UID: "209ad566-6ec9-44ba-9a0a-bc62cbfea583"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.243628 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/209ad566-6ec9-44ba-9a0a-bc62cbfea583-kube-api-access-qf67l" (OuterVolumeSpecName: "kube-api-access-qf67l") pod "209ad566-6ec9-44ba-9a0a-bc62cbfea583" (UID: "209ad566-6ec9-44ba-9a0a-bc62cbfea583"). InnerVolumeSpecName "kube-api-access-qf67l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.244092 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/209ad566-6ec9-44ba-9a0a-bc62cbfea583-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "209ad566-6ec9-44ba-9a0a-bc62cbfea583" (UID: "209ad566-6ec9-44ba-9a0a-bc62cbfea583"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.340288 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af8e7802-83b6-41c9-936c-b4837ecdd732-serving-cert\") pod \"route-controller-manager-c66dcf44f-f6gkh\" (UID: \"af8e7802-83b6-41c9-936c-b4837ecdd732\") " pod="openshift-route-controller-manager/route-controller-manager-c66dcf44f-f6gkh" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.340376 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af8e7802-83b6-41c9-936c-b4837ecdd732-config\") pod \"route-controller-manager-c66dcf44f-f6gkh\" (UID: \"af8e7802-83b6-41c9-936c-b4837ecdd732\") " pod="openshift-route-controller-manager/route-controller-manager-c66dcf44f-f6gkh" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.340402 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v4ww8\" (UniqueName: \"kubernetes.io/projected/af8e7802-83b6-41c9-936c-b4837ecdd732-kube-api-access-v4ww8\") pod \"route-controller-manager-c66dcf44f-f6gkh\" (UID: \"af8e7802-83b6-41c9-936c-b4837ecdd732\") " pod="openshift-route-controller-manager/route-controller-manager-c66dcf44f-f6gkh" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.340564 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af8e7802-83b6-41c9-936c-b4837ecdd732-client-ca\") pod \"route-controller-manager-c66dcf44f-f6gkh\" (UID: \"af8e7802-83b6-41c9-936c-b4837ecdd732\") " pod="openshift-route-controller-manager/route-controller-manager-c66dcf44f-f6gkh" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.340618 4894 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/209ad566-6ec9-44ba-9a0a-bc62cbfea583-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.340657 4894 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/209ad566-6ec9-44ba-9a0a-bc62cbfea583-client-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.340671 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/209ad566-6ec9-44ba-9a0a-bc62cbfea583-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.340684 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qf67l\" (UniqueName: \"kubernetes.io/projected/209ad566-6ec9-44ba-9a0a-bc62cbfea583-kube-api-access-qf67l\") on node \"crc\" DevicePath \"\"" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.341682 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/af8e7802-83b6-41c9-936c-b4837ecdd732-client-ca\") pod \"route-controller-manager-c66dcf44f-f6gkh\" (UID: \"af8e7802-83b6-41c9-936c-b4837ecdd732\") " pod="openshift-route-controller-manager/route-controller-manager-c66dcf44f-f6gkh" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.342178 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/af8e7802-83b6-41c9-936c-b4837ecdd732-config\") pod \"route-controller-manager-c66dcf44f-f6gkh\" (UID: \"af8e7802-83b6-41c9-936c-b4837ecdd732\") " pod="openshift-route-controller-manager/route-controller-manager-c66dcf44f-f6gkh" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.344697 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/af8e7802-83b6-41c9-936c-b4837ecdd732-serving-cert\") pod \"route-controller-manager-c66dcf44f-f6gkh\" (UID: \"af8e7802-83b6-41c9-936c-b4837ecdd732\") " pod="openshift-route-controller-manager/route-controller-manager-c66dcf44f-f6gkh" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.357370 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v4ww8\" (UniqueName: \"kubernetes.io/projected/af8e7802-83b6-41c9-936c-b4837ecdd732-kube-api-access-v4ww8\") pod \"route-controller-manager-c66dcf44f-f6gkh\" (UID: \"af8e7802-83b6-41c9-936c-b4837ecdd732\") " pod="openshift-route-controller-manager/route-controller-manager-c66dcf44f-f6gkh" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.424502 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-c66dcf44f-f6gkh" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.875022 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-c66dcf44f-f6gkh"] Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.879182 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.879520 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd" event={"ID":"209ad566-6ec9-44ba-9a0a-bc62cbfea583","Type":"ContainerDied","Data":"f7ab22717de4904698f011724e58dc882364726f4bdedc43e02adfa1684231a2"} Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.879598 4894 scope.go:117] "RemoveContainer" containerID="6a5a84367ea89e1aa00cd610ed60fa492af84bf3b3be6523d156055c59749b1a" Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.880896 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c66dcf44f-f6gkh" event={"ID":"af8e7802-83b6-41c9-936c-b4837ecdd732","Type":"ContainerStarted","Data":"97e1ae248029fe63d2543e39844134dba1a21071105eab2541da5cf95b0cb4f1"} Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.909747 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd"] Dec 09 15:37:50 crc kubenswrapper[4894]: I1209 15:37:50.912529 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7484d9ddcc-44hvd"] Dec 09 15:37:51 crc kubenswrapper[4894]: I1209 15:37:51.889340 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-c66dcf44f-f6gkh" event={"ID":"af8e7802-83b6-41c9-936c-b4837ecdd732","Type":"ContainerStarted","Data":"a820aa1a1fc7ff12dfde3e4a4eae5a1ec4b86a18b7331be97146d1602f2a2667"} Dec 09 15:37:51 crc kubenswrapper[4894]: I1209 15:37:51.889731 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-c66dcf44f-f6gkh" Dec 09 15:37:51 crc kubenswrapper[4894]: I1209 15:37:51.894817 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-c66dcf44f-f6gkh" Dec 09 15:37:51 crc kubenswrapper[4894]: I1209 15:37:51.908847 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-c66dcf44f-f6gkh" podStartSLOduration=4.908828547 podStartE2EDuration="4.908828547s" podCreationTimestamp="2025-12-09 15:37:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:37:51.905893596 +0000 UTC m=+366.225104275" watchObservedRunningTime="2025-12-09 15:37:51.908828547 +0000 UTC m=+366.228039216" Dec 09 15:37:52 crc kubenswrapper[4894]: I1209 15:37:52.113616 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="209ad566-6ec9-44ba-9a0a-bc62cbfea583" path="/var/lib/kubelet/pods/209ad566-6ec9-44ba-9a0a-bc62cbfea583/volumes" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.482600 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" podUID="569bec4c-7397-468e-ad4f-dd69ee0ed023" containerName="oauth-openshift" containerID="cri-o://4e6e6eff4898e2307b04ddaf3a855e22814c12e1d72d16d3c33d43290e05907f" gracePeriod=15 Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.840715 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.859098 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-ocp-branding-template\") pod \"569bec4c-7397-468e-ad4f-dd69ee0ed023\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.859182 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-service-ca\") pod \"569bec4c-7397-468e-ad4f-dd69ee0ed023\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.859246 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-cliconfig\") pod \"569bec4c-7397-468e-ad4f-dd69ee0ed023\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.859269 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d6br\" (UniqueName: \"kubernetes.io/projected/569bec4c-7397-468e-ad4f-dd69ee0ed023-kube-api-access-4d6br\") pod \"569bec4c-7397-468e-ad4f-dd69ee0ed023\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.859320 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/569bec4c-7397-468e-ad4f-dd69ee0ed023-audit-dir\") pod \"569bec4c-7397-468e-ad4f-dd69ee0ed023\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.859338 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-template-error\") pod \"569bec4c-7397-468e-ad4f-dd69ee0ed023\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.859356 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-router-certs\") pod \"569bec4c-7397-468e-ad4f-dd69ee0ed023\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.859394 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-template-provider-selection\") pod \"569bec4c-7397-468e-ad4f-dd69ee0ed023\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.859415 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-idp-0-file-data\") pod \"569bec4c-7397-468e-ad4f-dd69ee0ed023\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.859432 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-audit-policies\") pod \"569bec4c-7397-468e-ad4f-dd69ee0ed023\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.859507 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-serving-cert\") pod \"569bec4c-7397-468e-ad4f-dd69ee0ed023\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.859547 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-session\") pod \"569bec4c-7397-468e-ad4f-dd69ee0ed023\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.859572 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-template-login\") pod \"569bec4c-7397-468e-ad4f-dd69ee0ed023\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.859589 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-trusted-ca-bundle\") pod \"569bec4c-7397-468e-ad4f-dd69ee0ed023\" (UID: \"569bec4c-7397-468e-ad4f-dd69ee0ed023\") " Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.860625 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/569bec4c-7397-468e-ad4f-dd69ee0ed023-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "569bec4c-7397-468e-ad4f-dd69ee0ed023" (UID: "569bec4c-7397-468e-ad4f-dd69ee0ed023"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.860653 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "569bec4c-7397-468e-ad4f-dd69ee0ed023" (UID: "569bec4c-7397-468e-ad4f-dd69ee0ed023"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.861779 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "569bec4c-7397-468e-ad4f-dd69ee0ed023" (UID: "569bec4c-7397-468e-ad4f-dd69ee0ed023"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.861836 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "569bec4c-7397-468e-ad4f-dd69ee0ed023" (UID: "569bec4c-7397-468e-ad4f-dd69ee0ed023"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.862485 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "569bec4c-7397-468e-ad4f-dd69ee0ed023" (UID: "569bec4c-7397-468e-ad4f-dd69ee0ed023"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.866170 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "569bec4c-7397-468e-ad4f-dd69ee0ed023" (UID: "569bec4c-7397-468e-ad4f-dd69ee0ed023"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.869783 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "569bec4c-7397-468e-ad4f-dd69ee0ed023" (UID: "569bec4c-7397-468e-ad4f-dd69ee0ed023"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.871188 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "569bec4c-7397-468e-ad4f-dd69ee0ed023" (UID: "569bec4c-7397-468e-ad4f-dd69ee0ed023"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.871557 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "569bec4c-7397-468e-ad4f-dd69ee0ed023" (UID: "569bec4c-7397-468e-ad4f-dd69ee0ed023"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.871716 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "569bec4c-7397-468e-ad4f-dd69ee0ed023" (UID: "569bec4c-7397-468e-ad4f-dd69ee0ed023"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.871733 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/569bec4c-7397-468e-ad4f-dd69ee0ed023-kube-api-access-4d6br" (OuterVolumeSpecName: "kube-api-access-4d6br") pod "569bec4c-7397-468e-ad4f-dd69ee0ed023" (UID: "569bec4c-7397-468e-ad4f-dd69ee0ed023"). InnerVolumeSpecName "kube-api-access-4d6br". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.872176 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "569bec4c-7397-468e-ad4f-dd69ee0ed023" (UID: "569bec4c-7397-468e-ad4f-dd69ee0ed023"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.873946 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "569bec4c-7397-468e-ad4f-dd69ee0ed023" (UID: "569bec4c-7397-468e-ad4f-dd69ee0ed023"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.880776 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-586d5b9769-6hj5z"] Dec 09 15:37:58 crc kubenswrapper[4894]: E1209 15:37:58.881031 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="569bec4c-7397-468e-ad4f-dd69ee0ed023" containerName="oauth-openshift" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.881043 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="569bec4c-7397-468e-ad4f-dd69ee0ed023" containerName="oauth-openshift" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.881140 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="569bec4c-7397-468e-ad4f-dd69ee0ed023" containerName="oauth-openshift" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.881593 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.885107 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "569bec4c-7397-468e-ad4f-dd69ee0ed023" (UID: "569bec4c-7397-468e-ad4f-dd69ee0ed023"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.885769 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-586d5b9769-6hj5z"] Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.924909 4894 generic.go:334] "Generic (PLEG): container finished" podID="569bec4c-7397-468e-ad4f-dd69ee0ed023" containerID="4e6e6eff4898e2307b04ddaf3a855e22814c12e1d72d16d3c33d43290e05907f" exitCode=0 Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.924955 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" event={"ID":"569bec4c-7397-468e-ad4f-dd69ee0ed023","Type":"ContainerDied","Data":"4e6e6eff4898e2307b04ddaf3a855e22814c12e1d72d16d3c33d43290e05907f"} Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.924967 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.924990 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-g79bw" event={"ID":"569bec4c-7397-468e-ad4f-dd69ee0ed023","Type":"ContainerDied","Data":"e2bb18758cf504aa6fb0bb4f0e1e9df3b2409e25166b6ef20276127e03b6d511"} Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.925013 4894 scope.go:117] "RemoveContainer" containerID="4e6e6eff4898e2307b04ddaf3a855e22814c12e1d72d16d3c33d43290e05907f" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.940225 4894 scope.go:117] "RemoveContainer" containerID="4e6e6eff4898e2307b04ddaf3a855e22814c12e1d72d16d3c33d43290e05907f" Dec 09 15:37:58 crc kubenswrapper[4894]: E1209 15:37:58.940913 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4e6e6eff4898e2307b04ddaf3a855e22814c12e1d72d16d3c33d43290e05907f\": container with ID starting with 4e6e6eff4898e2307b04ddaf3a855e22814c12e1d72d16d3c33d43290e05907f not found: ID does not exist" containerID="4e6e6eff4898e2307b04ddaf3a855e22814c12e1d72d16d3c33d43290e05907f" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.940966 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4e6e6eff4898e2307b04ddaf3a855e22814c12e1d72d16d3c33d43290e05907f"} err="failed to get container status \"4e6e6eff4898e2307b04ddaf3a855e22814c12e1d72d16d3c33d43290e05907f\": rpc error: code = NotFound desc = could not find container \"4e6e6eff4898e2307b04ddaf3a855e22814c12e1d72d16d3c33d43290e05907f\": container with ID starting with 4e6e6eff4898e2307b04ddaf3a855e22814c12e1d72d16d3c33d43290e05907f not found: ID does not exist" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.957192 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-g79bw"] Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.960672 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-g79bw"] Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961171 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961216 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-system-router-certs\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961243 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961268 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-system-session\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961311 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961337 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961369 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/33be65d5-9e52-4f37-9a82-7ae670c7633b-audit-dir\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961395 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-user-template-error\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961422 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-user-template-login\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961481 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961508 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/33be65d5-9e52-4f37-9a82-7ae670c7633b-audit-policies\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961541 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-system-service-ca\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961569 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wqltw\" (UniqueName: \"kubernetes.io/projected/33be65d5-9e52-4f37-9a82-7ae670c7633b-kube-api-access-wqltw\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961646 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961706 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961741 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961754 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961769 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961783 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961795 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961808 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961819 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d6br\" (UniqueName: \"kubernetes.io/projected/569bec4c-7397-468e-ad4f-dd69ee0ed023-kube-api-access-4d6br\") on node \"crc\" DevicePath \"\"" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961831 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961843 4894 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/569bec4c-7397-468e-ad4f-dd69ee0ed023-audit-dir\") on node \"crc\" DevicePath \"\"" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961855 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961867 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961879 4894 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/569bec4c-7397-468e-ad4f-dd69ee0ed023-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:37:58 crc kubenswrapper[4894]: I1209 15:37:58.961888 4894 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/569bec4c-7397-468e-ad4f-dd69ee0ed023-audit-policies\") on node \"crc\" DevicePath \"\"" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.063605 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.064050 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.064155 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-system-router-certs\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.064229 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.064326 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-system-session\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.064423 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.064510 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.064600 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/33be65d5-9e52-4f37-9a82-7ae670c7633b-audit-dir\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.064700 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-user-template-error\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.064818 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-user-template-login\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.064954 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.065039 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/33be65d5-9e52-4f37-9a82-7ae670c7633b-audit-policies\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.065119 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-system-service-ca\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.065200 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wqltw\" (UniqueName: \"kubernetes.io/projected/33be65d5-9e52-4f37-9a82-7ae670c7633b-kube-api-access-wqltw\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.065697 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/33be65d5-9e52-4f37-9a82-7ae670c7633b-audit-policies\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.064731 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/33be65d5-9e52-4f37-9a82-7ae670c7633b-audit-dir\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.066059 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.066189 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-system-cliconfig\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.066774 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-system-service-ca\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.067708 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-system-router-certs\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.068506 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.068782 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-user-template-error\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.069102 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-system-session\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.069313 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.070018 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-user-template-login\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.070311 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.070546 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/33be65d5-9e52-4f37-9a82-7ae670c7633b-v4-0-config-system-serving-cert\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.083410 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wqltw\" (UniqueName: \"kubernetes.io/projected/33be65d5-9e52-4f37-9a82-7ae670c7633b-kube-api-access-wqltw\") pod \"oauth-openshift-586d5b9769-6hj5z\" (UID: \"33be65d5-9e52-4f37-9a82-7ae670c7633b\") " pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.224277 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.590033 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-586d5b9769-6hj5z"] Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.930861 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" event={"ID":"33be65d5-9e52-4f37-9a82-7ae670c7633b","Type":"ContainerStarted","Data":"e49025d7f1b2454aa791e650a92384e5108103e0fae7ea01cb7765d420fa79a7"} Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.931217 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.931236 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" event={"ID":"33be65d5-9e52-4f37-9a82-7ae670c7633b","Type":"ContainerStarted","Data":"0ff23e4630ed94785b3ad9d646dd7408a6b74f717b5bf939854e53eb68cf838e"} Dec 09 15:37:59 crc kubenswrapper[4894]: I1209 15:37:59.954178 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" podStartSLOduration=26.954157536 podStartE2EDuration="26.954157536s" podCreationTimestamp="2025-12-09 15:37:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:37:59.947107678 +0000 UTC m=+374.266318357" watchObservedRunningTime="2025-12-09 15:37:59.954157536 +0000 UTC m=+374.273368205" Dec 09 15:38:00 crc kubenswrapper[4894]: I1209 15:38:00.113257 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="569bec4c-7397-468e-ad4f-dd69ee0ed023" path="/var/lib/kubelet/pods/569bec4c-7397-468e-ad4f-dd69ee0ed023/volumes" Dec 09 15:38:00 crc kubenswrapper[4894]: I1209 15:38:00.537270 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-586d5b9769-6hj5z" Dec 09 15:38:07 crc kubenswrapper[4894]: I1209 15:38:07.982434 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" podUID="adab756d-1eb4-4837-a93a-346a38f83df9" containerName="registry" containerID="cri-o://c023c8982a0c0ad50ff5f66272882f7a5ffdc6ffbeba98d17209cd6023c76928" gracePeriod=30 Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.361452 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.475300 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"adab756d-1eb4-4837-a93a-346a38f83df9\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.475368 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/adab756d-1eb4-4837-a93a-346a38f83df9-registry-tls\") pod \"adab756d-1eb4-4837-a93a-346a38f83df9\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.475398 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/adab756d-1eb4-4837-a93a-346a38f83df9-trusted-ca\") pod \"adab756d-1eb4-4837-a93a-346a38f83df9\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.475446 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ppm42\" (UniqueName: \"kubernetes.io/projected/adab756d-1eb4-4837-a93a-346a38f83df9-kube-api-access-ppm42\") pod \"adab756d-1eb4-4837-a93a-346a38f83df9\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.475472 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/adab756d-1eb4-4837-a93a-346a38f83df9-registry-certificates\") pod \"adab756d-1eb4-4837-a93a-346a38f83df9\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.475542 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/adab756d-1eb4-4837-a93a-346a38f83df9-ca-trust-extracted\") pod \"adab756d-1eb4-4837-a93a-346a38f83df9\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.475572 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/adab756d-1eb4-4837-a93a-346a38f83df9-bound-sa-token\") pod \"adab756d-1eb4-4837-a93a-346a38f83df9\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.475596 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/adab756d-1eb4-4837-a93a-346a38f83df9-installation-pull-secrets\") pod \"adab756d-1eb4-4837-a93a-346a38f83df9\" (UID: \"adab756d-1eb4-4837-a93a-346a38f83df9\") " Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.477467 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adab756d-1eb4-4837-a93a-346a38f83df9-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "adab756d-1eb4-4837-a93a-346a38f83df9" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.477891 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/adab756d-1eb4-4837-a93a-346a38f83df9-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "adab756d-1eb4-4837-a93a-346a38f83df9" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.481602 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/adab756d-1eb4-4837-a93a-346a38f83df9-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "adab756d-1eb4-4837-a93a-346a38f83df9" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.481875 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adab756d-1eb4-4837-a93a-346a38f83df9-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "adab756d-1eb4-4837-a93a-346a38f83df9" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.482135 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adab756d-1eb4-4837-a93a-346a38f83df9-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "adab756d-1eb4-4837-a93a-346a38f83df9" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.482501 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/adab756d-1eb4-4837-a93a-346a38f83df9-kube-api-access-ppm42" (OuterVolumeSpecName: "kube-api-access-ppm42") pod "adab756d-1eb4-4837-a93a-346a38f83df9" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9"). InnerVolumeSpecName "kube-api-access-ppm42". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.488538 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "adab756d-1eb4-4837-a93a-346a38f83df9" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.493880 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/adab756d-1eb4-4837-a93a-346a38f83df9-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "adab756d-1eb4-4837-a93a-346a38f83df9" (UID: "adab756d-1eb4-4837-a93a-346a38f83df9"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.577092 4894 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/adab756d-1eb4-4837-a93a-346a38f83df9-bound-sa-token\") on node \"crc\" DevicePath \"\"" Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.577126 4894 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/adab756d-1eb4-4837-a93a-346a38f83df9-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.577140 4894 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/adab756d-1eb4-4837-a93a-346a38f83df9-registry-tls\") on node \"crc\" DevicePath \"\"" Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.577152 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/adab756d-1eb4-4837-a93a-346a38f83df9-trusted-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.577162 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ppm42\" (UniqueName: \"kubernetes.io/projected/adab756d-1eb4-4837-a93a-346a38f83df9-kube-api-access-ppm42\") on node \"crc\" DevicePath \"\"" Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.577175 4894 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/adab756d-1eb4-4837-a93a-346a38f83df9-registry-certificates\") on node \"crc\" DevicePath \"\"" Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.577185 4894 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/adab756d-1eb4-4837-a93a-346a38f83df9-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.984086 4894 generic.go:334] "Generic (PLEG): container finished" podID="adab756d-1eb4-4837-a93a-346a38f83df9" containerID="c023c8982a0c0ad50ff5f66272882f7a5ffdc6ffbeba98d17209cd6023c76928" exitCode=0 Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.984132 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" event={"ID":"adab756d-1eb4-4837-a93a-346a38f83df9","Type":"ContainerDied","Data":"c023c8982a0c0ad50ff5f66272882f7a5ffdc6ffbeba98d17209cd6023c76928"} Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.984172 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" event={"ID":"adab756d-1eb4-4837-a93a-346a38f83df9","Type":"ContainerDied","Data":"e1954edfeedc886a54a667e9bc01c139c7487849899b090c4d5e26c2b0e662c2"} Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.984192 4894 scope.go:117] "RemoveContainer" containerID="c023c8982a0c0ad50ff5f66272882f7a5ffdc6ffbeba98d17209cd6023c76928" Dec 09 15:38:08 crc kubenswrapper[4894]: I1209 15:38:08.984191 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-k8jmv" Dec 09 15:38:09 crc kubenswrapper[4894]: I1209 15:38:09.004026 4894 scope.go:117] "RemoveContainer" containerID="c023c8982a0c0ad50ff5f66272882f7a5ffdc6ffbeba98d17209cd6023c76928" Dec 09 15:38:09 crc kubenswrapper[4894]: E1209 15:38:09.004796 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c023c8982a0c0ad50ff5f66272882f7a5ffdc6ffbeba98d17209cd6023c76928\": container with ID starting with c023c8982a0c0ad50ff5f66272882f7a5ffdc6ffbeba98d17209cd6023c76928 not found: ID does not exist" containerID="c023c8982a0c0ad50ff5f66272882f7a5ffdc6ffbeba98d17209cd6023c76928" Dec 09 15:38:09 crc kubenswrapper[4894]: I1209 15:38:09.004849 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c023c8982a0c0ad50ff5f66272882f7a5ffdc6ffbeba98d17209cd6023c76928"} err="failed to get container status \"c023c8982a0c0ad50ff5f66272882f7a5ffdc6ffbeba98d17209cd6023c76928\": rpc error: code = NotFound desc = could not find container \"c023c8982a0c0ad50ff5f66272882f7a5ffdc6ffbeba98d17209cd6023c76928\": container with ID starting with c023c8982a0c0ad50ff5f66272882f7a5ffdc6ffbeba98d17209cd6023c76928 not found: ID does not exist" Dec 09 15:38:09 crc kubenswrapper[4894]: I1209 15:38:09.015006 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k8jmv"] Dec 09 15:38:09 crc kubenswrapper[4894]: I1209 15:38:09.021271 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-k8jmv"] Dec 09 15:38:10 crc kubenswrapper[4894]: I1209 15:38:10.112865 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="adab756d-1eb4-4837-a93a-346a38f83df9" path="/var/lib/kubelet/pods/adab756d-1eb4-4837-a93a-346a38f83df9/volumes" Dec 09 15:38:12 crc kubenswrapper[4894]: I1209 15:38:12.257522 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:38:12 crc kubenswrapper[4894]: I1209 15:38:12.258141 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:38:12 crc kubenswrapper[4894]: I1209 15:38:12.258189 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 15:38:12 crc kubenswrapper[4894]: I1209 15:38:12.258831 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cff2cb26767fe11bd813d02b6541ceb81064036337295e671ed7939ae0026edf"} pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 15:38:12 crc kubenswrapper[4894]: I1209 15:38:12.258916 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" containerID="cri-o://cff2cb26767fe11bd813d02b6541ceb81064036337295e671ed7939ae0026edf" gracePeriod=600 Dec 09 15:38:13 crc kubenswrapper[4894]: I1209 15:38:13.014062 4894 generic.go:334] "Generic (PLEG): container finished" podID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerID="cff2cb26767fe11bd813d02b6541ceb81064036337295e671ed7939ae0026edf" exitCode=0 Dec 09 15:38:13 crc kubenswrapper[4894]: I1209 15:38:13.014134 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerDied","Data":"cff2cb26767fe11bd813d02b6541ceb81064036337295e671ed7939ae0026edf"} Dec 09 15:38:13 crc kubenswrapper[4894]: I1209 15:38:13.014627 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerStarted","Data":"d6ebc3b4a88a254cb43621cbd5525aefadbbb8b324c54d247db4a86dd26c746f"} Dec 09 15:38:13 crc kubenswrapper[4894]: I1209 15:38:13.014677 4894 scope.go:117] "RemoveContainer" containerID="f169febaabd64ebb1a8a053a8fb50ce02c274b2e8ba1223d3e78c0edb317836f" Dec 09 15:40:12 crc kubenswrapper[4894]: I1209 15:40:12.257678 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:40:12 crc kubenswrapper[4894]: I1209 15:40:12.258256 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:40:42 crc kubenswrapper[4894]: I1209 15:40:42.257374 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:40:42 crc kubenswrapper[4894]: I1209 15:40:42.258002 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:41:12 crc kubenswrapper[4894]: I1209 15:41:12.256998 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:41:12 crc kubenswrapper[4894]: I1209 15:41:12.258839 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:41:12 crc kubenswrapper[4894]: I1209 15:41:12.258999 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 15:41:12 crc kubenswrapper[4894]: I1209 15:41:12.259757 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d6ebc3b4a88a254cb43621cbd5525aefadbbb8b324c54d247db4a86dd26c746f"} pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 15:41:12 crc kubenswrapper[4894]: I1209 15:41:12.259920 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" containerID="cri-o://d6ebc3b4a88a254cb43621cbd5525aefadbbb8b324c54d247db4a86dd26c746f" gracePeriod=600 Dec 09 15:41:12 crc kubenswrapper[4894]: I1209 15:41:12.992584 4894 generic.go:334] "Generic (PLEG): container finished" podID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerID="d6ebc3b4a88a254cb43621cbd5525aefadbbb8b324c54d247db4a86dd26c746f" exitCode=0 Dec 09 15:41:12 crc kubenswrapper[4894]: I1209 15:41:12.992727 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerDied","Data":"d6ebc3b4a88a254cb43621cbd5525aefadbbb8b324c54d247db4a86dd26c746f"} Dec 09 15:41:12 crc kubenswrapper[4894]: I1209 15:41:12.992980 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerStarted","Data":"bf0d8a25593f3059a7a7f1286d689a889938309e9429a43b4edb9ebe0cc7eecf"} Dec 09 15:41:12 crc kubenswrapper[4894]: I1209 15:41:12.993005 4894 scope.go:117] "RemoveContainer" containerID="cff2cb26767fe11bd813d02b6541ceb81064036337295e671ed7939ae0026edf" Dec 09 15:42:05 crc kubenswrapper[4894]: I1209 15:42:05.889663 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-kwc95"] Dec 09 15:42:05 crc kubenswrapper[4894]: E1209 15:42:05.890249 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="adab756d-1eb4-4837-a93a-346a38f83df9" containerName="registry" Dec 09 15:42:05 crc kubenswrapper[4894]: I1209 15:42:05.890260 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="adab756d-1eb4-4837-a93a-346a38f83df9" containerName="registry" Dec 09 15:42:05 crc kubenswrapper[4894]: I1209 15:42:05.890354 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="adab756d-1eb4-4837-a93a-346a38f83df9" containerName="registry" Dec 09 15:42:05 crc kubenswrapper[4894]: I1209 15:42:05.890801 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-kwc95" Dec 09 15:42:05 crc kubenswrapper[4894]: I1209 15:42:05.893684 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-2f27v"] Dec 09 15:42:05 crc kubenswrapper[4894]: I1209 15:42:05.893695 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Dec 09 15:42:05 crc kubenswrapper[4894]: I1209 15:42:05.893701 4894 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-g9ccp" Dec 09 15:42:05 crc kubenswrapper[4894]: I1209 15:42:05.893751 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Dec 09 15:42:05 crc kubenswrapper[4894]: I1209 15:42:05.894382 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-2f27v" Dec 09 15:42:05 crc kubenswrapper[4894]: I1209 15:42:05.899267 4894 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-tdczd" Dec 09 15:42:05 crc kubenswrapper[4894]: I1209 15:42:05.902246 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-kwc95"] Dec 09 15:42:05 crc kubenswrapper[4894]: I1209 15:42:05.905575 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-2f27v"] Dec 09 15:42:05 crc kubenswrapper[4894]: I1209 15:42:05.933082 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-k8tfx"] Dec 09 15:42:05 crc kubenswrapper[4894]: I1209 15:42:05.933779 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-k8tfx" Dec 09 15:42:05 crc kubenswrapper[4894]: I1209 15:42:05.936095 4894 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-tn9kk" Dec 09 15:42:05 crc kubenswrapper[4894]: I1209 15:42:05.946576 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-k8tfx"] Dec 09 15:42:06 crc kubenswrapper[4894]: I1209 15:42:06.079457 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8n2kj\" (UniqueName: \"kubernetes.io/projected/74585f72-f240-4bdd-8f23-01874e9551fe-kube-api-access-8n2kj\") pod \"cert-manager-5b446d88c5-kwc95\" (UID: \"74585f72-f240-4bdd-8f23-01874e9551fe\") " pod="cert-manager/cert-manager-5b446d88c5-kwc95" Dec 09 15:42:06 crc kubenswrapper[4894]: I1209 15:42:06.079551 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv9bz\" (UniqueName: \"kubernetes.io/projected/9881dac0-8703-4808-8e2d-45d2df155dcf-kube-api-access-fv9bz\") pod \"cert-manager-webhook-5655c58dd6-k8tfx\" (UID: \"9881dac0-8703-4808-8e2d-45d2df155dcf\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-k8tfx" Dec 09 15:42:06 crc kubenswrapper[4894]: I1209 15:42:06.079595 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f58f\" (UniqueName: \"kubernetes.io/projected/d7727746-e66c-4645-af8d-56efc6b4ed82-kube-api-access-8f58f\") pod \"cert-manager-cainjector-7f985d654d-2f27v\" (UID: \"d7727746-e66c-4645-af8d-56efc6b4ed82\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-2f27v" Dec 09 15:42:06 crc kubenswrapper[4894]: I1209 15:42:06.181298 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8n2kj\" (UniqueName: \"kubernetes.io/projected/74585f72-f240-4bdd-8f23-01874e9551fe-kube-api-access-8n2kj\") pod \"cert-manager-5b446d88c5-kwc95\" (UID: \"74585f72-f240-4bdd-8f23-01874e9551fe\") " pod="cert-manager/cert-manager-5b446d88c5-kwc95" Dec 09 15:42:06 crc kubenswrapper[4894]: I1209 15:42:06.181422 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv9bz\" (UniqueName: \"kubernetes.io/projected/9881dac0-8703-4808-8e2d-45d2df155dcf-kube-api-access-fv9bz\") pod \"cert-manager-webhook-5655c58dd6-k8tfx\" (UID: \"9881dac0-8703-4808-8e2d-45d2df155dcf\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-k8tfx" Dec 09 15:42:06 crc kubenswrapper[4894]: I1209 15:42:06.181734 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f58f\" (UniqueName: \"kubernetes.io/projected/d7727746-e66c-4645-af8d-56efc6b4ed82-kube-api-access-8f58f\") pod \"cert-manager-cainjector-7f985d654d-2f27v\" (UID: \"d7727746-e66c-4645-af8d-56efc6b4ed82\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-2f27v" Dec 09 15:42:06 crc kubenswrapper[4894]: I1209 15:42:06.201544 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8n2kj\" (UniqueName: \"kubernetes.io/projected/74585f72-f240-4bdd-8f23-01874e9551fe-kube-api-access-8n2kj\") pod \"cert-manager-5b446d88c5-kwc95\" (UID: \"74585f72-f240-4bdd-8f23-01874e9551fe\") " pod="cert-manager/cert-manager-5b446d88c5-kwc95" Dec 09 15:42:06 crc kubenswrapper[4894]: I1209 15:42:06.203222 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f58f\" (UniqueName: \"kubernetes.io/projected/d7727746-e66c-4645-af8d-56efc6b4ed82-kube-api-access-8f58f\") pod \"cert-manager-cainjector-7f985d654d-2f27v\" (UID: \"d7727746-e66c-4645-af8d-56efc6b4ed82\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-2f27v" Dec 09 15:42:06 crc kubenswrapper[4894]: I1209 15:42:06.204790 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv9bz\" (UniqueName: \"kubernetes.io/projected/9881dac0-8703-4808-8e2d-45d2df155dcf-kube-api-access-fv9bz\") pod \"cert-manager-webhook-5655c58dd6-k8tfx\" (UID: \"9881dac0-8703-4808-8e2d-45d2df155dcf\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-k8tfx" Dec 09 15:42:06 crc kubenswrapper[4894]: I1209 15:42:06.213382 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-kwc95" Dec 09 15:42:06 crc kubenswrapper[4894]: I1209 15:42:06.220216 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-2f27v" Dec 09 15:42:06 crc kubenswrapper[4894]: I1209 15:42:06.250215 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-k8tfx" Dec 09 15:42:06 crc kubenswrapper[4894]: I1209 15:42:06.416869 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-2f27v"] Dec 09 15:42:06 crc kubenswrapper[4894]: I1209 15:42:06.458246 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-kwc95"] Dec 09 15:42:06 crc kubenswrapper[4894]: W1209 15:42:06.459791 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7727746_e66c_4645_af8d_56efc6b4ed82.slice/crio-e50a968b2fd34d8282b89b5bdc6bfb613aad419770d7aa4e1a6c463c4d190e66 WatchSource:0}: Error finding container e50a968b2fd34d8282b89b5bdc6bfb613aad419770d7aa4e1a6c463c4d190e66: Status 404 returned error can't find the container with id e50a968b2fd34d8282b89b5bdc6bfb613aad419770d7aa4e1a6c463c4d190e66 Dec 09 15:42:06 crc kubenswrapper[4894]: I1209 15:42:06.462752 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 15:42:06 crc kubenswrapper[4894]: W1209 15:42:06.466740 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod74585f72_f240_4bdd_8f23_01874e9551fe.slice/crio-3c55e312e60a0add7cc7c4959e47d2b657824939639e03d9cdd24ef403fd7d67 WatchSource:0}: Error finding container 3c55e312e60a0add7cc7c4959e47d2b657824939639e03d9cdd24ef403fd7d67: Status 404 returned error can't find the container with id 3c55e312e60a0add7cc7c4959e47d2b657824939639e03d9cdd24ef403fd7d67 Dec 09 15:42:06 crc kubenswrapper[4894]: I1209 15:42:06.484555 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-k8tfx"] Dec 09 15:42:06 crc kubenswrapper[4894]: W1209 15:42:06.491477 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9881dac0_8703_4808_8e2d_45d2df155dcf.slice/crio-f83ff998a0b387ef8dc363a8c67c511d186d89888dab4388c1b7d8cd4cc33f60 WatchSource:0}: Error finding container f83ff998a0b387ef8dc363a8c67c511d186d89888dab4388c1b7d8cd4cc33f60: Status 404 returned error can't find the container with id f83ff998a0b387ef8dc363a8c67c511d186d89888dab4388c1b7d8cd4cc33f60 Dec 09 15:42:07 crc kubenswrapper[4894]: I1209 15:42:07.346843 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-kwc95" event={"ID":"74585f72-f240-4bdd-8f23-01874e9551fe","Type":"ContainerStarted","Data":"3c55e312e60a0add7cc7c4959e47d2b657824939639e03d9cdd24ef403fd7d67"} Dec 09 15:42:07 crc kubenswrapper[4894]: I1209 15:42:07.348912 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-k8tfx" event={"ID":"9881dac0-8703-4808-8e2d-45d2df155dcf","Type":"ContainerStarted","Data":"f83ff998a0b387ef8dc363a8c67c511d186d89888dab4388c1b7d8cd4cc33f60"} Dec 09 15:42:07 crc kubenswrapper[4894]: I1209 15:42:07.350024 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-2f27v" event={"ID":"d7727746-e66c-4645-af8d-56efc6b4ed82","Type":"ContainerStarted","Data":"e50a968b2fd34d8282b89b5bdc6bfb613aad419770d7aa4e1a6c463c4d190e66"} Dec 09 15:42:09 crc kubenswrapper[4894]: I1209 15:42:09.369517 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-2f27v" event={"ID":"d7727746-e66c-4645-af8d-56efc6b4ed82","Type":"ContainerStarted","Data":"6de513ee2da4140c35d6c8f64b629f4cafac1d76c845fb3451a26f98b97249d2"} Dec 09 15:42:09 crc kubenswrapper[4894]: I1209 15:42:09.396316 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-2f27v" podStartSLOduration=2.156261562 podStartE2EDuration="4.396261812s" podCreationTimestamp="2025-12-09 15:42:05 +0000 UTC" firstStartedPulling="2025-12-09 15:42:06.462531313 +0000 UTC m=+620.781741982" lastFinishedPulling="2025-12-09 15:42:08.702531563 +0000 UTC m=+623.021742232" observedRunningTime="2025-12-09 15:42:09.393846367 +0000 UTC m=+623.713057046" watchObservedRunningTime="2025-12-09 15:42:09.396261812 +0000 UTC m=+623.715472491" Dec 09 15:42:10 crc kubenswrapper[4894]: I1209 15:42:10.376710 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-kwc95" event={"ID":"74585f72-f240-4bdd-8f23-01874e9551fe","Type":"ContainerStarted","Data":"d1a33d24a0b24359bd9dddce836ce6ac4bde3e9b89ad2476d7e0f137d2dafe1e"} Dec 09 15:42:10 crc kubenswrapper[4894]: I1209 15:42:10.383549 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-k8tfx" event={"ID":"9881dac0-8703-4808-8e2d-45d2df155dcf","Type":"ContainerStarted","Data":"2a61989425e908edf1f9c99b2530a3864c80ab64302c8424be7a6721f5865510"} Dec 09 15:42:10 crc kubenswrapper[4894]: I1209 15:42:10.383671 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-k8tfx" Dec 09 15:42:10 crc kubenswrapper[4894]: I1209 15:42:10.393035 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-kwc95" podStartSLOduration=1.91179401 podStartE2EDuration="5.393014438s" podCreationTimestamp="2025-12-09 15:42:05 +0000 UTC" firstStartedPulling="2025-12-09 15:42:06.469109401 +0000 UTC m=+620.788320070" lastFinishedPulling="2025-12-09 15:42:09.950329829 +0000 UTC m=+624.269540498" observedRunningTime="2025-12-09 15:42:10.392097654 +0000 UTC m=+624.711308363" watchObservedRunningTime="2025-12-09 15:42:10.393014438 +0000 UTC m=+624.712225107" Dec 09 15:42:10 crc kubenswrapper[4894]: I1209 15:42:10.407577 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-k8tfx" podStartSLOduration=2.005625343 podStartE2EDuration="5.407556723s" podCreationTimestamp="2025-12-09 15:42:05 +0000 UTC" firstStartedPulling="2025-12-09 15:42:06.492846674 +0000 UTC m=+620.812057343" lastFinishedPulling="2025-12-09 15:42:09.894778054 +0000 UTC m=+624.213988723" observedRunningTime="2025-12-09 15:42:10.405927488 +0000 UTC m=+624.725138157" watchObservedRunningTime="2025-12-09 15:42:10.407556723 +0000 UTC m=+624.726767392" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.232549 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-s5scr"] Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.233791 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovn-controller" containerID="cri-o://56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8" gracePeriod=30 Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.234302 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="sbdb" containerID="cri-o://8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0" gracePeriod=30 Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.234358 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed" gracePeriod=30 Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.234425 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="kube-rbac-proxy-node" containerID="cri-o://91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8" gracePeriod=30 Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.234499 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovn-acl-logging" containerID="cri-o://cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90" gracePeriod=30 Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.234514 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="northd" containerID="cri-o://860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102" gracePeriod=30 Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.234469 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="nbdb" containerID="cri-o://9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249" gracePeriod=30 Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.253856 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-k8tfx" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.278415 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovnkube-controller" containerID="cri-o://f27d1750c5bae1fad612cd4ef9f97c183ef4d7f6b2f9421f3b39e6fe5871a557" gracePeriod=30 Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.429440 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s5scr_97036c6f-dce3-444b-85e6-8476c16613c9/ovnkube-controller/3.log" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.438683 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s5scr_97036c6f-dce3-444b-85e6-8476c16613c9/ovn-acl-logging/0.log" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.439196 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s5scr_97036c6f-dce3-444b-85e6-8476c16613c9/ovn-controller/0.log" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.439825 4894 generic.go:334] "Generic (PLEG): container finished" podID="97036c6f-dce3-444b-85e6-8476c16613c9" containerID="f27d1750c5bae1fad612cd4ef9f97c183ef4d7f6b2f9421f3b39e6fe5871a557" exitCode=0 Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.439858 4894 generic.go:334] "Generic (PLEG): container finished" podID="97036c6f-dce3-444b-85e6-8476c16613c9" containerID="860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102" exitCode=0 Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.439869 4894 generic.go:334] "Generic (PLEG): container finished" podID="97036c6f-dce3-444b-85e6-8476c16613c9" containerID="4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed" exitCode=0 Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.439882 4894 generic.go:334] "Generic (PLEG): container finished" podID="97036c6f-dce3-444b-85e6-8476c16613c9" containerID="91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8" exitCode=0 Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.439894 4894 generic.go:334] "Generic (PLEG): container finished" podID="97036c6f-dce3-444b-85e6-8476c16613c9" containerID="cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90" exitCode=143 Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.439922 4894 generic.go:334] "Generic (PLEG): container finished" podID="97036c6f-dce3-444b-85e6-8476c16613c9" containerID="56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8" exitCode=143 Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.439977 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerDied","Data":"f27d1750c5bae1fad612cd4ef9f97c183ef4d7f6b2f9421f3b39e6fe5871a557"} Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.440012 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerDied","Data":"860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102"} Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.440027 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerDied","Data":"4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed"} Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.440040 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerDied","Data":"91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8"} Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.440053 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerDied","Data":"cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90"} Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.440065 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerDied","Data":"56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8"} Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.440085 4894 scope.go:117] "RemoveContainer" containerID="876913572e03e323189c646a2ea4359c30b652499f20ab2c876828c0fcc5ee36" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.443883 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qt56s_97d690f0-ac29-46c9-ad4a-b92bd0a30772/kube-multus/2.log" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.444315 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qt56s_97d690f0-ac29-46c9-ad4a-b92bd0a30772/kube-multus/1.log" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.444349 4894 generic.go:334] "Generic (PLEG): container finished" podID="97d690f0-ac29-46c9-ad4a-b92bd0a30772" containerID="1b46c71319fd2c655bf490b2b8ffaf6b6a6559171f0dc8a1d30985db9e3312de" exitCode=2 Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.444371 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qt56s" event={"ID":"97d690f0-ac29-46c9-ad4a-b92bd0a30772","Type":"ContainerDied","Data":"1b46c71319fd2c655bf490b2b8ffaf6b6a6559171f0dc8a1d30985db9e3312de"} Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.444732 4894 scope.go:117] "RemoveContainer" containerID="1b46c71319fd2c655bf490b2b8ffaf6b6a6559171f0dc8a1d30985db9e3312de" Dec 09 15:42:16 crc kubenswrapper[4894]: E1209 15:42:16.445031 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-qt56s_openshift-multus(97d690f0-ac29-46c9-ad4a-b92bd0a30772)\"" pod="openshift-multus/multus-qt56s" podUID="97d690f0-ac29-46c9-ad4a-b92bd0a30772" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.505619 4894 scope.go:117] "RemoveContainer" containerID="0cfd57bf783aa0e666bbe98fa7fc1eb2cda69cbd538b72fbb95491efd23e2037" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.535195 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s5scr_97036c6f-dce3-444b-85e6-8476c16613c9/ovn-acl-logging/0.log" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.535686 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s5scr_97036c6f-dce3-444b-85e6-8476c16613c9/ovn-controller/0.log" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.536091 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.540799 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-slash\") pod \"97036c6f-dce3-444b-85e6-8476c16613c9\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.540826 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-kubelet\") pod \"97036c6f-dce3-444b-85e6-8476c16613c9\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.540845 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-run-netns\") pod \"97036c6f-dce3-444b-85e6-8476c16613c9\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.540866 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-var-lib-openvswitch\") pod \"97036c6f-dce3-444b-85e6-8476c16613c9\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.540872 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "97036c6f-dce3-444b-85e6-8476c16613c9" (UID: "97036c6f-dce3-444b-85e6-8476c16613c9"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.540888 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/97036c6f-dce3-444b-85e6-8476c16613c9-env-overrides\") pod \"97036c6f-dce3-444b-85e6-8476c16613c9\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.540900 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "97036c6f-dce3-444b-85e6-8476c16613c9" (UID: "97036c6f-dce3-444b-85e6-8476c16613c9"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.540919 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "97036c6f-dce3-444b-85e6-8476c16613c9" (UID: "97036c6f-dce3-444b-85e6-8476c16613c9"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.540913 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-slash" (OuterVolumeSpecName: "host-slash") pod "97036c6f-dce3-444b-85e6-8476c16613c9" (UID: "97036c6f-dce3-444b-85e6-8476c16613c9"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.541398 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97036c6f-dce3-444b-85e6-8476c16613c9-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "97036c6f-dce3-444b-85e6-8476c16613c9" (UID: "97036c6f-dce3-444b-85e6-8476c16613c9"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.541628 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-cni-bin\") pod \"97036c6f-dce3-444b-85e6-8476c16613c9\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.541691 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-etc-openvswitch\") pod \"97036c6f-dce3-444b-85e6-8476c16613c9\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.541723 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "97036c6f-dce3-444b-85e6-8476c16613c9" (UID: "97036c6f-dce3-444b-85e6-8476c16613c9"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.541797 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-run-ovn\") pod \"97036c6f-dce3-444b-85e6-8476c16613c9\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.541809 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "97036c6f-dce3-444b-85e6-8476c16613c9" (UID: "97036c6f-dce3-444b-85e6-8476c16613c9"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.541821 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"97036c6f-dce3-444b-85e6-8476c16613c9\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.541850 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/97036c6f-dce3-444b-85e6-8476c16613c9-ovn-node-metrics-cert\") pod \"97036c6f-dce3-444b-85e6-8476c16613c9\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.541877 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/97036c6f-dce3-444b-85e6-8476c16613c9-ovnkube-config\") pod \"97036c6f-dce3-444b-85e6-8476c16613c9\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.541881 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "97036c6f-dce3-444b-85e6-8476c16613c9" (UID: "97036c6f-dce3-444b-85e6-8476c16613c9"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.541903 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-log-socket\") pod \"97036c6f-dce3-444b-85e6-8476c16613c9\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.541904 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "97036c6f-dce3-444b-85e6-8476c16613c9" (UID: "97036c6f-dce3-444b-85e6-8476c16613c9"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.541932 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-node-log\") pod \"97036c6f-dce3-444b-85e6-8476c16613c9\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.541946 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-run-openvswitch\") pod \"97036c6f-dce3-444b-85e6-8476c16613c9\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.541969 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-systemd-units\") pod \"97036c6f-dce3-444b-85e6-8476c16613c9\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.541977 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-log-socket" (OuterVolumeSpecName: "log-socket") pod "97036c6f-dce3-444b-85e6-8476c16613c9" (UID: "97036c6f-dce3-444b-85e6-8476c16613c9"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.541996 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-cni-netd\") pod \"97036c6f-dce3-444b-85e6-8476c16613c9\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.542003 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "97036c6f-dce3-444b-85e6-8476c16613c9" (UID: "97036c6f-dce3-444b-85e6-8476c16613c9"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.542012 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-run-systemd\") pod \"97036c6f-dce3-444b-85e6-8476c16613c9\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.542025 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-node-log" (OuterVolumeSpecName: "node-log") pod "97036c6f-dce3-444b-85e6-8476c16613c9" (UID: "97036c6f-dce3-444b-85e6-8476c16613c9"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.542037 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5l7x\" (UniqueName: \"kubernetes.io/projected/97036c6f-dce3-444b-85e6-8476c16613c9-kube-api-access-d5l7x\") pod \"97036c6f-dce3-444b-85e6-8476c16613c9\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.542050 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "97036c6f-dce3-444b-85e6-8476c16613c9" (UID: "97036c6f-dce3-444b-85e6-8476c16613c9"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.542059 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-run-ovn-kubernetes\") pod \"97036c6f-dce3-444b-85e6-8476c16613c9\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.542114 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "97036c6f-dce3-444b-85e6-8476c16613c9" (UID: "97036c6f-dce3-444b-85e6-8476c16613c9"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.542119 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/97036c6f-dce3-444b-85e6-8476c16613c9-ovnkube-script-lib\") pod \"97036c6f-dce3-444b-85e6-8476c16613c9\" (UID: \"97036c6f-dce3-444b-85e6-8476c16613c9\") " Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.542615 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97036c6f-dce3-444b-85e6-8476c16613c9-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "97036c6f-dce3-444b-85e6-8476c16613c9" (UID: "97036c6f-dce3-444b-85e6-8476c16613c9"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.542666 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "97036c6f-dce3-444b-85e6-8476c16613c9" (UID: "97036c6f-dce3-444b-85e6-8476c16613c9"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.542794 4894 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-log-socket\") on node \"crc\" DevicePath \"\"" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.542814 4894 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-node-log\") on node \"crc\" DevicePath \"\"" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.542828 4894 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-run-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.542840 4894 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-systemd-units\") on node \"crc\" DevicePath \"\"" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.542852 4894 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-cni-netd\") on node \"crc\" DevicePath \"\"" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.542856 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/97036c6f-dce3-444b-85e6-8476c16613c9-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "97036c6f-dce3-444b-85e6-8476c16613c9" (UID: "97036c6f-dce3-444b-85e6-8476c16613c9"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.542864 4894 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.542931 4894 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/97036c6f-dce3-444b-85e6-8476c16613c9-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.542947 4894 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-slash\") on node \"crc\" DevicePath \"\"" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.542981 4894 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-kubelet\") on node \"crc\" DevicePath \"\"" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.542992 4894 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-run-netns\") on node \"crc\" DevicePath \"\"" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.543001 4894 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.543010 4894 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/97036c6f-dce3-444b-85e6-8476c16613c9-env-overrides\") on node \"crc\" DevicePath \"\"" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.543019 4894 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-cni-bin\") on node \"crc\" DevicePath \"\"" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.543027 4894 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.543036 4894 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.543071 4894 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.548036 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/97036c6f-dce3-444b-85e6-8476c16613c9-kube-api-access-d5l7x" (OuterVolumeSpecName: "kube-api-access-d5l7x") pod "97036c6f-dce3-444b-85e6-8476c16613c9" (UID: "97036c6f-dce3-444b-85e6-8476c16613c9"). InnerVolumeSpecName "kube-api-access-d5l7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.548488 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/97036c6f-dce3-444b-85e6-8476c16613c9-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "97036c6f-dce3-444b-85e6-8476c16613c9" (UID: "97036c6f-dce3-444b-85e6-8476c16613c9"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.556521 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "97036c6f-dce3-444b-85e6-8476c16613c9" (UID: "97036c6f-dce3-444b-85e6-8476c16613c9"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.583168 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-h5bkp"] Dec 09 15:42:16 crc kubenswrapper[4894]: E1209 15:42:16.583768 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="nbdb" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.583864 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="nbdb" Dec 09 15:42:16 crc kubenswrapper[4894]: E1209 15:42:16.583936 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovn-acl-logging" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.584013 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovn-acl-logging" Dec 09 15:42:16 crc kubenswrapper[4894]: E1209 15:42:16.584129 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="kube-rbac-proxy-ovn-metrics" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.584244 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="kube-rbac-proxy-ovn-metrics" Dec 09 15:42:16 crc kubenswrapper[4894]: E1209 15:42:16.584325 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="kube-rbac-proxy-node" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.584401 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="kube-rbac-proxy-node" Dec 09 15:42:16 crc kubenswrapper[4894]: E1209 15:42:16.584468 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovnkube-controller" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.584537 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovnkube-controller" Dec 09 15:42:16 crc kubenswrapper[4894]: E1209 15:42:16.584591 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="sbdb" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.584684 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="sbdb" Dec 09 15:42:16 crc kubenswrapper[4894]: E1209 15:42:16.584754 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovnkube-controller" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.584828 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovnkube-controller" Dec 09 15:42:16 crc kubenswrapper[4894]: E1209 15:42:16.584897 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="kubecfg-setup" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.584952 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="kubecfg-setup" Dec 09 15:42:16 crc kubenswrapper[4894]: E1209 15:42:16.585188 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovn-controller" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.585254 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovn-controller" Dec 09 15:42:16 crc kubenswrapper[4894]: E1209 15:42:16.585322 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovnkube-controller" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.585386 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovnkube-controller" Dec 09 15:42:16 crc kubenswrapper[4894]: E1209 15:42:16.585435 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovnkube-controller" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.585507 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovnkube-controller" Dec 09 15:42:16 crc kubenswrapper[4894]: E1209 15:42:16.585581 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="northd" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.585663 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="northd" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.585813 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovn-acl-logging" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.585872 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="kube-rbac-proxy-node" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.585921 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovnkube-controller" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.585964 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovnkube-controller" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.586014 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovn-controller" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.586152 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="nbdb" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.586225 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="kube-rbac-proxy-ovn-metrics" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.586342 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovnkube-controller" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.586411 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="northd" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.586457 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovnkube-controller" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.586509 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="sbdb" Dec 09 15:42:16 crc kubenswrapper[4894]: E1209 15:42:16.586676 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovnkube-controller" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.586738 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovnkube-controller" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.586946 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" containerName="ovnkube-controller" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.600125 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644104 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-run-ovn\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644161 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644198 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-ovn-node-metrics-cert\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644249 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h4szb\" (UniqueName: \"kubernetes.io/projected/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-kube-api-access-h4szb\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644271 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-env-overrides\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644294 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-host-cni-bin\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644328 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-run-openvswitch\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644346 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-ovnkube-config\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644371 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-host-kubelet\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644389 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-node-log\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644406 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-host-run-ovn-kubernetes\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644425 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-etc-openvswitch\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644441 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-systemd-units\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644464 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-var-lib-openvswitch\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644519 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-log-socket\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644557 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-run-systemd\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644577 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-host-slash\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644610 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-ovnkube-script-lib\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644626 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-host-run-netns\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644694 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-host-cni-netd\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644761 4894 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/97036c6f-dce3-444b-85e6-8476c16613c9-run-systemd\") on node \"crc\" DevicePath \"\"" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644777 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5l7x\" (UniqueName: \"kubernetes.io/projected/97036c6f-dce3-444b-85e6-8476c16613c9-kube-api-access-d5l7x\") on node \"crc\" DevicePath \"\"" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644947 4894 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/97036c6f-dce3-444b-85e6-8476c16613c9-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.644957 4894 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/97036c6f-dce3-444b-85e6-8476c16613c9-ovnkube-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.746861 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-host-kubelet\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.746938 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-node-log\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.746983 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-host-run-ovn-kubernetes\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747017 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-host-kubelet\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747031 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-etc-openvswitch\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747096 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-systemd-units\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747101 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-etc-openvswitch\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747130 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-var-lib-openvswitch\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747101 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-node-log\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747097 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-host-run-ovn-kubernetes\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747163 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-systemd-units\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747192 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-var-lib-openvswitch\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747178 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-log-socket\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747217 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-log-socket\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747251 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-run-systemd\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747272 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-host-slash\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747324 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-ovnkube-script-lib\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747342 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-host-run-netns\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747360 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-host-slash\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747405 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-host-cni-netd\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747431 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-run-ovn\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747464 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747498 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-ovn-node-metrics-cert\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747317 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-run-systemd\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747520 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h4szb\" (UniqueName: \"kubernetes.io/projected/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-kube-api-access-h4szb\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747540 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-env-overrides\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747506 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-host-run-netns\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747558 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-run-ovn\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747597 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-host-cni-bin\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747614 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-host-cni-netd\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747556 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-host-cni-bin\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747623 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747909 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-run-openvswitch\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747951 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-run-openvswitch\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.747967 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-ovnkube-config\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.748286 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-env-overrides\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.748313 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-ovnkube-script-lib\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.748554 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-ovnkube-config\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.755305 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-ovn-node-metrics-cert\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.765542 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h4szb\" (UniqueName: \"kubernetes.io/projected/da5c41b8-d58d-4a4f-8a45-af13097ec5a5-kube-api-access-h4szb\") pod \"ovnkube-node-h5bkp\" (UID: \"da5c41b8-d58d-4a4f-8a45-af13097ec5a5\") " pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: I1209 15:42:16.926363 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:16 crc kubenswrapper[4894]: W1209 15:42:16.958020 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podda5c41b8_d58d_4a4f_8a45_af13097ec5a5.slice/crio-7048bc46555e03175a0e5637690d5d72452762aded4d57e2d81a191adfda8ea4 WatchSource:0}: Error finding container 7048bc46555e03175a0e5637690d5d72452762aded4d57e2d81a191adfda8ea4: Status 404 returned error can't find the container with id 7048bc46555e03175a0e5637690d5d72452762aded4d57e2d81a191adfda8ea4 Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.452705 4894 generic.go:334] "Generic (PLEG): container finished" podID="da5c41b8-d58d-4a4f-8a45-af13097ec5a5" containerID="1ee736d2908bd1c9c0c7e9ed8d59bd632ea71d48ba9e8b68a9bee7c5053e4223" exitCode=0 Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.452744 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" event={"ID":"da5c41b8-d58d-4a4f-8a45-af13097ec5a5","Type":"ContainerDied","Data":"1ee736d2908bd1c9c0c7e9ed8d59bd632ea71d48ba9e8b68a9bee7c5053e4223"} Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.452798 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" event={"ID":"da5c41b8-d58d-4a4f-8a45-af13097ec5a5","Type":"ContainerStarted","Data":"7048bc46555e03175a0e5637690d5d72452762aded4d57e2d81a191adfda8ea4"} Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.459879 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s5scr_97036c6f-dce3-444b-85e6-8476c16613c9/ovn-acl-logging/0.log" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.460480 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-s5scr_97036c6f-dce3-444b-85e6-8476c16613c9/ovn-controller/0.log" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.460971 4894 generic.go:334] "Generic (PLEG): container finished" podID="97036c6f-dce3-444b-85e6-8476c16613c9" containerID="8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0" exitCode=0 Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.460994 4894 generic.go:334] "Generic (PLEG): container finished" podID="97036c6f-dce3-444b-85e6-8476c16613c9" containerID="9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249" exitCode=0 Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.461016 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerDied","Data":"8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0"} Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.461059 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerDied","Data":"9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249"} Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.461078 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" event={"ID":"97036c6f-dce3-444b-85e6-8476c16613c9","Type":"ContainerDied","Data":"9cd60c8a7e5113f8c18495faaa522a4339748e2680b99424222e32596d12b619"} Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.461090 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-s5scr" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.461078 4894 scope.go:117] "RemoveContainer" containerID="f27d1750c5bae1fad612cd4ef9f97c183ef4d7f6b2f9421f3b39e6fe5871a557" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.464217 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qt56s_97d690f0-ac29-46c9-ad4a-b92bd0a30772/kube-multus/2.log" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.486869 4894 scope.go:117] "RemoveContainer" containerID="8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.513040 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-s5scr"] Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.515050 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-s5scr"] Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.543589 4894 scope.go:117] "RemoveContainer" containerID="9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.575880 4894 scope.go:117] "RemoveContainer" containerID="860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.590003 4894 scope.go:117] "RemoveContainer" containerID="4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.606464 4894 scope.go:117] "RemoveContainer" containerID="91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.633159 4894 scope.go:117] "RemoveContainer" containerID="cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.649110 4894 scope.go:117] "RemoveContainer" containerID="56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.675122 4894 scope.go:117] "RemoveContainer" containerID="993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.694341 4894 scope.go:117] "RemoveContainer" containerID="f27d1750c5bae1fad612cd4ef9f97c183ef4d7f6b2f9421f3b39e6fe5871a557" Dec 09 15:42:17 crc kubenswrapper[4894]: E1209 15:42:17.694765 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f27d1750c5bae1fad612cd4ef9f97c183ef4d7f6b2f9421f3b39e6fe5871a557\": container with ID starting with f27d1750c5bae1fad612cd4ef9f97c183ef4d7f6b2f9421f3b39e6fe5871a557 not found: ID does not exist" containerID="f27d1750c5bae1fad612cd4ef9f97c183ef4d7f6b2f9421f3b39e6fe5871a557" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.694814 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f27d1750c5bae1fad612cd4ef9f97c183ef4d7f6b2f9421f3b39e6fe5871a557"} err="failed to get container status \"f27d1750c5bae1fad612cd4ef9f97c183ef4d7f6b2f9421f3b39e6fe5871a557\": rpc error: code = NotFound desc = could not find container \"f27d1750c5bae1fad612cd4ef9f97c183ef4d7f6b2f9421f3b39e6fe5871a557\": container with ID starting with f27d1750c5bae1fad612cd4ef9f97c183ef4d7f6b2f9421f3b39e6fe5871a557 not found: ID does not exist" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.694840 4894 scope.go:117] "RemoveContainer" containerID="8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0" Dec 09 15:42:17 crc kubenswrapper[4894]: E1209 15:42:17.695201 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\": container with ID starting with 8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0 not found: ID does not exist" containerID="8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.695226 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0"} err="failed to get container status \"8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\": rpc error: code = NotFound desc = could not find container \"8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\": container with ID starting with 8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0 not found: ID does not exist" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.695242 4894 scope.go:117] "RemoveContainer" containerID="9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249" Dec 09 15:42:17 crc kubenswrapper[4894]: E1209 15:42:17.695546 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\": container with ID starting with 9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249 not found: ID does not exist" containerID="9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.695570 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249"} err="failed to get container status \"9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\": rpc error: code = NotFound desc = could not find container \"9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\": container with ID starting with 9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249 not found: ID does not exist" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.695586 4894 scope.go:117] "RemoveContainer" containerID="860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102" Dec 09 15:42:17 crc kubenswrapper[4894]: E1209 15:42:17.696052 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\": container with ID starting with 860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102 not found: ID does not exist" containerID="860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.696078 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102"} err="failed to get container status \"860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\": rpc error: code = NotFound desc = could not find container \"860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\": container with ID starting with 860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102 not found: ID does not exist" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.696097 4894 scope.go:117] "RemoveContainer" containerID="4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed" Dec 09 15:42:17 crc kubenswrapper[4894]: E1209 15:42:17.696329 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\": container with ID starting with 4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed not found: ID does not exist" containerID="4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.696352 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed"} err="failed to get container status \"4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\": rpc error: code = NotFound desc = could not find container \"4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\": container with ID starting with 4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed not found: ID does not exist" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.696368 4894 scope.go:117] "RemoveContainer" containerID="91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8" Dec 09 15:42:17 crc kubenswrapper[4894]: E1209 15:42:17.696649 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\": container with ID starting with 91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8 not found: ID does not exist" containerID="91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.696674 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8"} err="failed to get container status \"91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\": rpc error: code = NotFound desc = could not find container \"91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\": container with ID starting with 91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8 not found: ID does not exist" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.696692 4894 scope.go:117] "RemoveContainer" containerID="cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90" Dec 09 15:42:17 crc kubenswrapper[4894]: E1209 15:42:17.696988 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\": container with ID starting with cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90 not found: ID does not exist" containerID="cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.697013 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90"} err="failed to get container status \"cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\": rpc error: code = NotFound desc = could not find container \"cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\": container with ID starting with cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90 not found: ID does not exist" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.697032 4894 scope.go:117] "RemoveContainer" containerID="56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8" Dec 09 15:42:17 crc kubenswrapper[4894]: E1209 15:42:17.697423 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\": container with ID starting with 56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8 not found: ID does not exist" containerID="56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.697445 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8"} err="failed to get container status \"56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\": rpc error: code = NotFound desc = could not find container \"56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\": container with ID starting with 56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8 not found: ID does not exist" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.697463 4894 scope.go:117] "RemoveContainer" containerID="993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034" Dec 09 15:42:17 crc kubenswrapper[4894]: E1209 15:42:17.697736 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\": container with ID starting with 993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034 not found: ID does not exist" containerID="993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.697758 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034"} err="failed to get container status \"993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\": rpc error: code = NotFound desc = could not find container \"993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\": container with ID starting with 993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034 not found: ID does not exist" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.697776 4894 scope.go:117] "RemoveContainer" containerID="f27d1750c5bae1fad612cd4ef9f97c183ef4d7f6b2f9421f3b39e6fe5871a557" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.698145 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f27d1750c5bae1fad612cd4ef9f97c183ef4d7f6b2f9421f3b39e6fe5871a557"} err="failed to get container status \"f27d1750c5bae1fad612cd4ef9f97c183ef4d7f6b2f9421f3b39e6fe5871a557\": rpc error: code = NotFound desc = could not find container \"f27d1750c5bae1fad612cd4ef9f97c183ef4d7f6b2f9421f3b39e6fe5871a557\": container with ID starting with f27d1750c5bae1fad612cd4ef9f97c183ef4d7f6b2f9421f3b39e6fe5871a557 not found: ID does not exist" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.698174 4894 scope.go:117] "RemoveContainer" containerID="8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.698427 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0"} err="failed to get container status \"8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\": rpc error: code = NotFound desc = could not find container \"8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0\": container with ID starting with 8b2c729c2b529df9cecb574f2a7429890d653c8bbb1fb26c3d099a3cf12b1ab0 not found: ID does not exist" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.698454 4894 scope.go:117] "RemoveContainer" containerID="9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.698668 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249"} err="failed to get container status \"9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\": rpc error: code = NotFound desc = could not find container \"9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249\": container with ID starting with 9ee20760b5f6bb37a9edd484968a18d731aba05f597829ea8de5204130e9f249 not found: ID does not exist" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.698691 4894 scope.go:117] "RemoveContainer" containerID="860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.698937 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102"} err="failed to get container status \"860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\": rpc error: code = NotFound desc = could not find container \"860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102\": container with ID starting with 860e4141114006fac2a09fa077710bcff49e0cb63747f9e1894ec72531354102 not found: ID does not exist" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.698965 4894 scope.go:117] "RemoveContainer" containerID="4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.699179 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed"} err="failed to get container status \"4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\": rpc error: code = NotFound desc = could not find container \"4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed\": container with ID starting with 4ad89137aeae81921585cbeb1f8cfbc1675c0250ba35cfcb75c185b43ad909ed not found: ID does not exist" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.699202 4894 scope.go:117] "RemoveContainer" containerID="91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.699413 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8"} err="failed to get container status \"91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\": rpc error: code = NotFound desc = could not find container \"91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8\": container with ID starting with 91e2a971d4ada0a8abe9236c411b865833accb7821dfb35a96c86dff82e6c9c8 not found: ID does not exist" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.699437 4894 scope.go:117] "RemoveContainer" containerID="cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.699855 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90"} err="failed to get container status \"cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\": rpc error: code = NotFound desc = could not find container \"cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90\": container with ID starting with cf5b4c3cba24036a1910443d25e15cf9dd00970a6a07f5380b2f5c006fc10b90 not found: ID does not exist" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.699877 4894 scope.go:117] "RemoveContainer" containerID="56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.700112 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8"} err="failed to get container status \"56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\": rpc error: code = NotFound desc = could not find container \"56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8\": container with ID starting with 56bd1ff45f1e00a6f87518790b64810b412c5069b8d2b1ea0cc73ad95c5b1ff8 not found: ID does not exist" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.700133 4894 scope.go:117] "RemoveContainer" containerID="993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034" Dec 09 15:42:17 crc kubenswrapper[4894]: I1209 15:42:17.700542 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034"} err="failed to get container status \"993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\": rpc error: code = NotFound desc = could not find container \"993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034\": container with ID starting with 993a6de6e49245422e2774f9bb58addbc2cf790359dc0b4d750659480226d034 not found: ID does not exist" Dec 09 15:42:18 crc kubenswrapper[4894]: I1209 15:42:18.117072 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="97036c6f-dce3-444b-85e6-8476c16613c9" path="/var/lib/kubelet/pods/97036c6f-dce3-444b-85e6-8476c16613c9/volumes" Dec 09 15:42:18 crc kubenswrapper[4894]: I1209 15:42:18.471813 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" event={"ID":"da5c41b8-d58d-4a4f-8a45-af13097ec5a5","Type":"ContainerStarted","Data":"ba832faa0491bf670689a23df760e956918b4356004d5a9675a8a04a373f9852"} Dec 09 15:42:18 crc kubenswrapper[4894]: I1209 15:42:18.471857 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" event={"ID":"da5c41b8-d58d-4a4f-8a45-af13097ec5a5","Type":"ContainerStarted","Data":"326d0dd371c0d5b0b564b41b179994f9e6a9f05937717ea52f9f245874702367"} Dec 09 15:42:18 crc kubenswrapper[4894]: I1209 15:42:18.471871 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" event={"ID":"da5c41b8-d58d-4a4f-8a45-af13097ec5a5","Type":"ContainerStarted","Data":"d762555b302d739bddcfe11597183ad1f15a91a391684aff77fc38b1521dc62d"} Dec 09 15:42:18 crc kubenswrapper[4894]: I1209 15:42:18.471883 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" event={"ID":"da5c41b8-d58d-4a4f-8a45-af13097ec5a5","Type":"ContainerStarted","Data":"59aa368beff454a8de3dc046ed52140c14e187e61efa9e90d891fc5b6fa19316"} Dec 09 15:42:18 crc kubenswrapper[4894]: I1209 15:42:18.471893 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" event={"ID":"da5c41b8-d58d-4a4f-8a45-af13097ec5a5","Type":"ContainerStarted","Data":"f03505ed03f4e5e2230c2e14088f982a852ef6c22b37e663a34082219d375e19"} Dec 09 15:42:18 crc kubenswrapper[4894]: I1209 15:42:18.471904 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" event={"ID":"da5c41b8-d58d-4a4f-8a45-af13097ec5a5","Type":"ContainerStarted","Data":"a9b8a6dbcfd14c7b0f589ac2206ced2a435d8161598d588822a09263ff4cdd0b"} Dec 09 15:42:20 crc kubenswrapper[4894]: I1209 15:42:20.490353 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" event={"ID":"da5c41b8-d58d-4a4f-8a45-af13097ec5a5","Type":"ContainerStarted","Data":"22d0ac4bfc7a2212d1b706ffb7b885cec4de4dab1584016c81cb1619a59a396f"} Dec 09 15:42:23 crc kubenswrapper[4894]: I1209 15:42:23.514546 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" event={"ID":"da5c41b8-d58d-4a4f-8a45-af13097ec5a5","Type":"ContainerStarted","Data":"19982418f52c2cb9a9f12461b633546aa3a9083ef292b7608dc2faae76c22680"} Dec 09 15:42:23 crc kubenswrapper[4894]: I1209 15:42:23.514952 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:23 crc kubenswrapper[4894]: I1209 15:42:23.514969 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:23 crc kubenswrapper[4894]: I1209 15:42:23.514979 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:23 crc kubenswrapper[4894]: I1209 15:42:23.552287 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" podStartSLOduration=7.552268093 podStartE2EDuration="7.552268093s" podCreationTimestamp="2025-12-09 15:42:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:42:23.547160104 +0000 UTC m=+637.866370773" watchObservedRunningTime="2025-12-09 15:42:23.552268093 +0000 UTC m=+637.871478772" Dec 09 15:42:23 crc kubenswrapper[4894]: I1209 15:42:23.556415 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:23 crc kubenswrapper[4894]: I1209 15:42:23.559260 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:42:30 crc kubenswrapper[4894]: I1209 15:42:30.106842 4894 scope.go:117] "RemoveContainer" containerID="1b46c71319fd2c655bf490b2b8ffaf6b6a6559171f0dc8a1d30985db9e3312de" Dec 09 15:42:30 crc kubenswrapper[4894]: E1209 15:42:30.112356 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-qt56s_openshift-multus(97d690f0-ac29-46c9-ad4a-b92bd0a30772)\"" pod="openshift-multus/multus-qt56s" podUID="97d690f0-ac29-46c9-ad4a-b92bd0a30772" Dec 09 15:42:45 crc kubenswrapper[4894]: I1209 15:42:45.106617 4894 scope.go:117] "RemoveContainer" containerID="1b46c71319fd2c655bf490b2b8ffaf6b6a6559171f0dc8a1d30985db9e3312de" Dec 09 15:42:46 crc kubenswrapper[4894]: I1209 15:42:46.685355 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-qt56s_97d690f0-ac29-46c9-ad4a-b92bd0a30772/kube-multus/2.log" Dec 09 15:42:46 crc kubenswrapper[4894]: I1209 15:42:46.686012 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-qt56s" event={"ID":"97d690f0-ac29-46c9-ad4a-b92bd0a30772","Type":"ContainerStarted","Data":"68a2c48036bb5681d7196280e054fcd6702c42e109b2afcdd6d648e47d6f50df"} Dec 09 15:42:46 crc kubenswrapper[4894]: I1209 15:42:46.945987 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-h5bkp" Dec 09 15:43:01 crc kubenswrapper[4894]: I1209 15:43:01.498911 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq"] Dec 09 15:43:01 crc kubenswrapper[4894]: I1209 15:43:01.501076 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq" Dec 09 15:43:01 crc kubenswrapper[4894]: I1209 15:43:01.511500 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 09 15:43:01 crc kubenswrapper[4894]: I1209 15:43:01.518741 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq"] Dec 09 15:43:01 crc kubenswrapper[4894]: I1209 15:43:01.661420 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtxfp\" (UniqueName: \"kubernetes.io/projected/f961acc2-dd50-4607-8ff5-d06a51bddf97-kube-api-access-mtxfp\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq\" (UID: \"f961acc2-dd50-4607-8ff5-d06a51bddf97\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq" Dec 09 15:43:01 crc kubenswrapper[4894]: I1209 15:43:01.661575 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f961acc2-dd50-4607-8ff5-d06a51bddf97-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq\" (UID: \"f961acc2-dd50-4607-8ff5-d06a51bddf97\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq" Dec 09 15:43:01 crc kubenswrapper[4894]: I1209 15:43:01.661786 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f961acc2-dd50-4607-8ff5-d06a51bddf97-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq\" (UID: \"f961acc2-dd50-4607-8ff5-d06a51bddf97\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq" Dec 09 15:43:01 crc kubenswrapper[4894]: I1209 15:43:01.763432 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f961acc2-dd50-4607-8ff5-d06a51bddf97-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq\" (UID: \"f961acc2-dd50-4607-8ff5-d06a51bddf97\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq" Dec 09 15:43:01 crc kubenswrapper[4894]: I1209 15:43:01.763562 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f961acc2-dd50-4607-8ff5-d06a51bddf97-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq\" (UID: \"f961acc2-dd50-4607-8ff5-d06a51bddf97\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq" Dec 09 15:43:01 crc kubenswrapper[4894]: I1209 15:43:01.763610 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtxfp\" (UniqueName: \"kubernetes.io/projected/f961acc2-dd50-4607-8ff5-d06a51bddf97-kube-api-access-mtxfp\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq\" (UID: \"f961acc2-dd50-4607-8ff5-d06a51bddf97\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq" Dec 09 15:43:01 crc kubenswrapper[4894]: I1209 15:43:01.764809 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f961acc2-dd50-4607-8ff5-d06a51bddf97-bundle\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq\" (UID: \"f961acc2-dd50-4607-8ff5-d06a51bddf97\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq" Dec 09 15:43:01 crc kubenswrapper[4894]: I1209 15:43:01.765237 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f961acc2-dd50-4607-8ff5-d06a51bddf97-util\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq\" (UID: \"f961acc2-dd50-4607-8ff5-d06a51bddf97\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq" Dec 09 15:43:01 crc kubenswrapper[4894]: I1209 15:43:01.790017 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtxfp\" (UniqueName: \"kubernetes.io/projected/f961acc2-dd50-4607-8ff5-d06a51bddf97-kube-api-access-mtxfp\") pod \"5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq\" (UID: \"f961acc2-dd50-4607-8ff5-d06a51bddf97\") " pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq" Dec 09 15:43:01 crc kubenswrapper[4894]: I1209 15:43:01.830829 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq" Dec 09 15:43:02 crc kubenswrapper[4894]: I1209 15:43:02.053324 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq"] Dec 09 15:43:02 crc kubenswrapper[4894]: I1209 15:43:02.769971 4894 generic.go:334] "Generic (PLEG): container finished" podID="f961acc2-dd50-4607-8ff5-d06a51bddf97" containerID="9abfaa4ef0ee79325e4e4e265f47e10d9fbcc93063c94ce44ecd38b87e04d3d9" exitCode=0 Dec 09 15:43:02 crc kubenswrapper[4894]: I1209 15:43:02.770045 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq" event={"ID":"f961acc2-dd50-4607-8ff5-d06a51bddf97","Type":"ContainerDied","Data":"9abfaa4ef0ee79325e4e4e265f47e10d9fbcc93063c94ce44ecd38b87e04d3d9"} Dec 09 15:43:02 crc kubenswrapper[4894]: I1209 15:43:02.770403 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq" event={"ID":"f961acc2-dd50-4607-8ff5-d06a51bddf97","Type":"ContainerStarted","Data":"2b5192497112d78f4de8efcbd51a3a6e613f90c642e6ae0ed47112b4bd9eb242"} Dec 09 15:43:04 crc kubenswrapper[4894]: I1209 15:43:04.780435 4894 generic.go:334] "Generic (PLEG): container finished" podID="f961acc2-dd50-4607-8ff5-d06a51bddf97" containerID="eaca2716526febfb46b14ebdd2b6010596aa4c9ced83959448794fae43f21b8a" exitCode=0 Dec 09 15:43:04 crc kubenswrapper[4894]: I1209 15:43:04.780519 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq" event={"ID":"f961acc2-dd50-4607-8ff5-d06a51bddf97","Type":"ContainerDied","Data":"eaca2716526febfb46b14ebdd2b6010596aa4c9ced83959448794fae43f21b8a"} Dec 09 15:43:05 crc kubenswrapper[4894]: I1209 15:43:05.788894 4894 generic.go:334] "Generic (PLEG): container finished" podID="f961acc2-dd50-4607-8ff5-d06a51bddf97" containerID="f2d2211b51d059eacae143ff547f797e0e1fbce5dbedebc2abcb42279529f5e5" exitCode=0 Dec 09 15:43:05 crc kubenswrapper[4894]: I1209 15:43:05.788968 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq" event={"ID":"f961acc2-dd50-4607-8ff5-d06a51bddf97","Type":"ContainerDied","Data":"f2d2211b51d059eacae143ff547f797e0e1fbce5dbedebc2abcb42279529f5e5"} Dec 09 15:43:07 crc kubenswrapper[4894]: I1209 15:43:07.041945 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq" Dec 09 15:43:07 crc kubenswrapper[4894]: I1209 15:43:07.100888 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f961acc2-dd50-4607-8ff5-d06a51bddf97-bundle\") pod \"f961acc2-dd50-4607-8ff5-d06a51bddf97\" (UID: \"f961acc2-dd50-4607-8ff5-d06a51bddf97\") " Dec 09 15:43:07 crc kubenswrapper[4894]: I1209 15:43:07.100974 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f961acc2-dd50-4607-8ff5-d06a51bddf97-util\") pod \"f961acc2-dd50-4607-8ff5-d06a51bddf97\" (UID: \"f961acc2-dd50-4607-8ff5-d06a51bddf97\") " Dec 09 15:43:07 crc kubenswrapper[4894]: I1209 15:43:07.101126 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtxfp\" (UniqueName: \"kubernetes.io/projected/f961acc2-dd50-4607-8ff5-d06a51bddf97-kube-api-access-mtxfp\") pod \"f961acc2-dd50-4607-8ff5-d06a51bddf97\" (UID: \"f961acc2-dd50-4607-8ff5-d06a51bddf97\") " Dec 09 15:43:07 crc kubenswrapper[4894]: I1209 15:43:07.101969 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f961acc2-dd50-4607-8ff5-d06a51bddf97-bundle" (OuterVolumeSpecName: "bundle") pod "f961acc2-dd50-4607-8ff5-d06a51bddf97" (UID: "f961acc2-dd50-4607-8ff5-d06a51bddf97"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:43:07 crc kubenswrapper[4894]: I1209 15:43:07.113876 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f961acc2-dd50-4607-8ff5-d06a51bddf97-kube-api-access-mtxfp" (OuterVolumeSpecName: "kube-api-access-mtxfp") pod "f961acc2-dd50-4607-8ff5-d06a51bddf97" (UID: "f961acc2-dd50-4607-8ff5-d06a51bddf97"). InnerVolumeSpecName "kube-api-access-mtxfp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:43:07 crc kubenswrapper[4894]: I1209 15:43:07.116684 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f961acc2-dd50-4607-8ff5-d06a51bddf97-util" (OuterVolumeSpecName: "util") pod "f961acc2-dd50-4607-8ff5-d06a51bddf97" (UID: "f961acc2-dd50-4607-8ff5-d06a51bddf97"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:43:07 crc kubenswrapper[4894]: I1209 15:43:07.203501 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtxfp\" (UniqueName: \"kubernetes.io/projected/f961acc2-dd50-4607-8ff5-d06a51bddf97-kube-api-access-mtxfp\") on node \"crc\" DevicePath \"\"" Dec 09 15:43:07 crc kubenswrapper[4894]: I1209 15:43:07.203537 4894 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/f961acc2-dd50-4607-8ff5-d06a51bddf97-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:43:07 crc kubenswrapper[4894]: I1209 15:43:07.203547 4894 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/f961acc2-dd50-4607-8ff5-d06a51bddf97-util\") on node \"crc\" DevicePath \"\"" Dec 09 15:43:07 crc kubenswrapper[4894]: I1209 15:43:07.802849 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq" event={"ID":"f961acc2-dd50-4607-8ff5-d06a51bddf97","Type":"ContainerDied","Data":"2b5192497112d78f4de8efcbd51a3a6e613f90c642e6ae0ed47112b4bd9eb242"} Dec 09 15:43:07 crc kubenswrapper[4894]: I1209 15:43:07.802895 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b5192497112d78f4de8efcbd51a3a6e613f90c642e6ae0ed47112b4bd9eb242" Dec 09 15:43:07 crc kubenswrapper[4894]: I1209 15:43:07.802967 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq" Dec 09 15:43:10 crc kubenswrapper[4894]: I1209 15:43:10.119397 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-5ctgb"] Dec 09 15:43:10 crc kubenswrapper[4894]: E1209 15:43:10.119954 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f961acc2-dd50-4607-8ff5-d06a51bddf97" containerName="pull" Dec 09 15:43:10 crc kubenswrapper[4894]: I1209 15:43:10.119967 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f961acc2-dd50-4607-8ff5-d06a51bddf97" containerName="pull" Dec 09 15:43:10 crc kubenswrapper[4894]: E1209 15:43:10.119983 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f961acc2-dd50-4607-8ff5-d06a51bddf97" containerName="extract" Dec 09 15:43:10 crc kubenswrapper[4894]: I1209 15:43:10.119991 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f961acc2-dd50-4607-8ff5-d06a51bddf97" containerName="extract" Dec 09 15:43:10 crc kubenswrapper[4894]: E1209 15:43:10.119998 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f961acc2-dd50-4607-8ff5-d06a51bddf97" containerName="util" Dec 09 15:43:10 crc kubenswrapper[4894]: I1209 15:43:10.120004 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f961acc2-dd50-4607-8ff5-d06a51bddf97" containerName="util" Dec 09 15:43:10 crc kubenswrapper[4894]: I1209 15:43:10.120106 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f961acc2-dd50-4607-8ff5-d06a51bddf97" containerName="extract" Dec 09 15:43:10 crc kubenswrapper[4894]: I1209 15:43:10.120503 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5ctgb" Dec 09 15:43:10 crc kubenswrapper[4894]: I1209 15:43:10.123373 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Dec 09 15:43:10 crc kubenswrapper[4894]: I1209 15:43:10.123409 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-g4gdw" Dec 09 15:43:10 crc kubenswrapper[4894]: I1209 15:43:10.123807 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Dec 09 15:43:10 crc kubenswrapper[4894]: I1209 15:43:10.138337 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-5ctgb"] Dec 09 15:43:10 crc kubenswrapper[4894]: I1209 15:43:10.146072 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b5zr5\" (UniqueName: \"kubernetes.io/projected/a6aa74db-0590-4c30-ae7a-164d405f3944-kube-api-access-b5zr5\") pod \"nmstate-operator-5b5b58f5c8-5ctgb\" (UID: \"a6aa74db-0590-4c30-ae7a-164d405f3944\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5ctgb" Dec 09 15:43:10 crc kubenswrapper[4894]: I1209 15:43:10.246819 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b5zr5\" (UniqueName: \"kubernetes.io/projected/a6aa74db-0590-4c30-ae7a-164d405f3944-kube-api-access-b5zr5\") pod \"nmstate-operator-5b5b58f5c8-5ctgb\" (UID: \"a6aa74db-0590-4c30-ae7a-164d405f3944\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5ctgb" Dec 09 15:43:10 crc kubenswrapper[4894]: I1209 15:43:10.269023 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b5zr5\" (UniqueName: \"kubernetes.io/projected/a6aa74db-0590-4c30-ae7a-164d405f3944-kube-api-access-b5zr5\") pod \"nmstate-operator-5b5b58f5c8-5ctgb\" (UID: \"a6aa74db-0590-4c30-ae7a-164d405f3944\") " pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5ctgb" Dec 09 15:43:10 crc kubenswrapper[4894]: I1209 15:43:10.434213 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5ctgb" Dec 09 15:43:10 crc kubenswrapper[4894]: I1209 15:43:10.843266 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-5b5b58f5c8-5ctgb"] Dec 09 15:43:11 crc kubenswrapper[4894]: I1209 15:43:11.825610 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5ctgb" event={"ID":"a6aa74db-0590-4c30-ae7a-164d405f3944","Type":"ContainerStarted","Data":"788115cd316951363a9dc88b16cab6644a826befe2150484262346fcc520f361"} Dec 09 15:43:12 crc kubenswrapper[4894]: I1209 15:43:12.257686 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:43:12 crc kubenswrapper[4894]: I1209 15:43:12.257745 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:43:13 crc kubenswrapper[4894]: I1209 15:43:13.835940 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5ctgb" event={"ID":"a6aa74db-0590-4c30-ae7a-164d405f3944","Type":"ContainerStarted","Data":"9f4b0b6404a3b1e9255a2f6803809527b47e5d5a372be355bdc8e41e692bf19e"} Dec 09 15:43:13 crc kubenswrapper[4894]: I1209 15:43:13.855042 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-5b5b58f5c8-5ctgb" podStartSLOduration=1.721060995 podStartE2EDuration="3.85502375s" podCreationTimestamp="2025-12-09 15:43:10 +0000 UTC" firstStartedPulling="2025-12-09 15:43:10.851255165 +0000 UTC m=+685.170465834" lastFinishedPulling="2025-12-09 15:43:12.98521791 +0000 UTC m=+687.304428589" observedRunningTime="2025-12-09 15:43:13.852380989 +0000 UTC m=+688.171591688" watchObservedRunningTime="2025-12-09 15:43:13.85502375 +0000 UTC m=+688.174234429" Dec 09 15:43:14 crc kubenswrapper[4894]: I1209 15:43:14.835877 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-tclzs"] Dec 09 15:43:14 crc kubenswrapper[4894]: I1209 15:43:14.837281 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-tclzs" Dec 09 15:43:14 crc kubenswrapper[4894]: I1209 15:43:14.839541 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-cqlwd" Dec 09 15:43:14 crc kubenswrapper[4894]: I1209 15:43:14.843329 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5zmgn"] Dec 09 15:43:14 crc kubenswrapper[4894]: I1209 15:43:14.844141 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5zmgn" Dec 09 15:43:14 crc kubenswrapper[4894]: I1209 15:43:14.855508 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Dec 09 15:43:14 crc kubenswrapper[4894]: I1209 15:43:14.864549 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-tclzs"] Dec 09 15:43:14 crc kubenswrapper[4894]: I1209 15:43:14.893355 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5zmgn"] Dec 09 15:43:14 crc kubenswrapper[4894]: I1209 15:43:14.901120 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-4vf4p"] Dec 09 15:43:14 crc kubenswrapper[4894]: I1209 15:43:14.901871 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-4vf4p" Dec 09 15:43:14 crc kubenswrapper[4894]: I1209 15:43:14.907890 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/34e69c07-d85c-4a34-afe7-287647f01876-ovs-socket\") pod \"nmstate-handler-4vf4p\" (UID: \"34e69c07-d85c-4a34-afe7-287647f01876\") " pod="openshift-nmstate/nmstate-handler-4vf4p" Dec 09 15:43:14 crc kubenswrapper[4894]: I1209 15:43:14.907952 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/2cbfef40-9c05-4039-b1db-d47e119ab8da-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-5zmgn\" (UID: \"2cbfef40-9c05-4039-b1db-d47e119ab8da\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5zmgn" Dec 09 15:43:14 crc kubenswrapper[4894]: I1209 15:43:14.908009 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/34e69c07-d85c-4a34-afe7-287647f01876-dbus-socket\") pod \"nmstate-handler-4vf4p\" (UID: \"34e69c07-d85c-4a34-afe7-287647f01876\") " pod="openshift-nmstate/nmstate-handler-4vf4p" Dec 09 15:43:14 crc kubenswrapper[4894]: I1209 15:43:14.908032 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hgqj\" (UniqueName: \"kubernetes.io/projected/34e69c07-d85c-4a34-afe7-287647f01876-kube-api-access-5hgqj\") pod \"nmstate-handler-4vf4p\" (UID: \"34e69c07-d85c-4a34-afe7-287647f01876\") " pod="openshift-nmstate/nmstate-handler-4vf4p" Dec 09 15:43:14 crc kubenswrapper[4894]: I1209 15:43:14.908138 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4jrjq\" (UniqueName: \"kubernetes.io/projected/81fe83b5-37f9-499e-af80-602c302d905c-kube-api-access-4jrjq\") pod \"nmstate-metrics-7f946cbc9-tclzs\" (UID: \"81fe83b5-37f9-499e-af80-602c302d905c\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-tclzs" Dec 09 15:43:14 crc kubenswrapper[4894]: I1209 15:43:14.908270 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/34e69c07-d85c-4a34-afe7-287647f01876-nmstate-lock\") pod \"nmstate-handler-4vf4p\" (UID: \"34e69c07-d85c-4a34-afe7-287647f01876\") " pod="openshift-nmstate/nmstate-handler-4vf4p" Dec 09 15:43:14 crc kubenswrapper[4894]: I1209 15:43:14.908304 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2jt9\" (UniqueName: \"kubernetes.io/projected/2cbfef40-9c05-4039-b1db-d47e119ab8da-kube-api-access-c2jt9\") pod \"nmstate-webhook-5f6d4c5ccb-5zmgn\" (UID: \"2cbfef40-9c05-4039-b1db-d47e119ab8da\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5zmgn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.009207 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/34e69c07-d85c-4a34-afe7-287647f01876-ovs-socket\") pod \"nmstate-handler-4vf4p\" (UID: \"34e69c07-d85c-4a34-afe7-287647f01876\") " pod="openshift-nmstate/nmstate-handler-4vf4p" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.009277 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/2cbfef40-9c05-4039-b1db-d47e119ab8da-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-5zmgn\" (UID: \"2cbfef40-9c05-4039-b1db-d47e119ab8da\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5zmgn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.009307 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/34e69c07-d85c-4a34-afe7-287647f01876-dbus-socket\") pod \"nmstate-handler-4vf4p\" (UID: \"34e69c07-d85c-4a34-afe7-287647f01876\") " pod="openshift-nmstate/nmstate-handler-4vf4p" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.009331 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hgqj\" (UniqueName: \"kubernetes.io/projected/34e69c07-d85c-4a34-afe7-287647f01876-kube-api-access-5hgqj\") pod \"nmstate-handler-4vf4p\" (UID: \"34e69c07-d85c-4a34-afe7-287647f01876\") " pod="openshift-nmstate/nmstate-handler-4vf4p" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.009368 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4jrjq\" (UniqueName: \"kubernetes.io/projected/81fe83b5-37f9-499e-af80-602c302d905c-kube-api-access-4jrjq\") pod \"nmstate-metrics-7f946cbc9-tclzs\" (UID: \"81fe83b5-37f9-499e-af80-602c302d905c\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-tclzs" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.009462 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/34e69c07-d85c-4a34-afe7-287647f01876-nmstate-lock\") pod \"nmstate-handler-4vf4p\" (UID: \"34e69c07-d85c-4a34-afe7-287647f01876\") " pod="openshift-nmstate/nmstate-handler-4vf4p" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.009487 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2jt9\" (UniqueName: \"kubernetes.io/projected/2cbfef40-9c05-4039-b1db-d47e119ab8da-kube-api-access-c2jt9\") pod \"nmstate-webhook-5f6d4c5ccb-5zmgn\" (UID: \"2cbfef40-9c05-4039-b1db-d47e119ab8da\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5zmgn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.010128 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/34e69c07-d85c-4a34-afe7-287647f01876-ovs-socket\") pod \"nmstate-handler-4vf4p\" (UID: \"34e69c07-d85c-4a34-afe7-287647f01876\") " pod="openshift-nmstate/nmstate-handler-4vf4p" Dec 09 15:43:15 crc kubenswrapper[4894]: E1209 15:43:15.010214 4894 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Dec 09 15:43:15 crc kubenswrapper[4894]: E1209 15:43:15.010266 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2cbfef40-9c05-4039-b1db-d47e119ab8da-tls-key-pair podName:2cbfef40-9c05-4039-b1db-d47e119ab8da nodeName:}" failed. No retries permitted until 2025-12-09 15:43:15.510246161 +0000 UTC m=+689.829456830 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/2cbfef40-9c05-4039-b1db-d47e119ab8da-tls-key-pair") pod "nmstate-webhook-5f6d4c5ccb-5zmgn" (UID: "2cbfef40-9c05-4039-b1db-d47e119ab8da") : secret "openshift-nmstate-webhook" not found Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.010728 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/34e69c07-d85c-4a34-afe7-287647f01876-dbus-socket\") pod \"nmstate-handler-4vf4p\" (UID: \"34e69c07-d85c-4a34-afe7-287647f01876\") " pod="openshift-nmstate/nmstate-handler-4vf4p" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.011086 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/34e69c07-d85c-4a34-afe7-287647f01876-nmstate-lock\") pod \"nmstate-handler-4vf4p\" (UID: \"34e69c07-d85c-4a34-afe7-287647f01876\") " pod="openshift-nmstate/nmstate-handler-4vf4p" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.037590 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4jrjq\" (UniqueName: \"kubernetes.io/projected/81fe83b5-37f9-499e-af80-602c302d905c-kube-api-access-4jrjq\") pod \"nmstate-metrics-7f946cbc9-tclzs\" (UID: \"81fe83b5-37f9-499e-af80-602c302d905c\") " pod="openshift-nmstate/nmstate-metrics-7f946cbc9-tclzs" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.038568 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2jt9\" (UniqueName: \"kubernetes.io/projected/2cbfef40-9c05-4039-b1db-d47e119ab8da-kube-api-access-c2jt9\") pod \"nmstate-webhook-5f6d4c5ccb-5zmgn\" (UID: \"2cbfef40-9c05-4039-b1db-d47e119ab8da\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5zmgn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.071050 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8vs59"] Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.071898 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8vs59" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.074178 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hgqj\" (UniqueName: \"kubernetes.io/projected/34e69c07-d85c-4a34-afe7-287647f01876-kube-api-access-5hgqj\") pod \"nmstate-handler-4vf4p\" (UID: \"34e69c07-d85c-4a34-afe7-287647f01876\") " pod="openshift-nmstate/nmstate-handler-4vf4p" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.075049 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.075086 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-vc7f9" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.075476 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.080759 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8vs59"] Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.111533 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5bk4\" (UniqueName: \"kubernetes.io/projected/61dcd6dd-61fc-4b15-8c34-b7b27775f1cd-kube-api-access-d5bk4\") pod \"nmstate-console-plugin-7fbb5f6569-8vs59\" (UID: \"61dcd6dd-61fc-4b15-8c34-b7b27775f1cd\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8vs59" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.111786 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/61dcd6dd-61fc-4b15-8c34-b7b27775f1cd-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-8vs59\" (UID: \"61dcd6dd-61fc-4b15-8c34-b7b27775f1cd\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8vs59" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.111906 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/61dcd6dd-61fc-4b15-8c34-b7b27775f1cd-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-8vs59\" (UID: \"61dcd6dd-61fc-4b15-8c34-b7b27775f1cd\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8vs59" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.154824 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-tclzs" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.212630 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5bk4\" (UniqueName: \"kubernetes.io/projected/61dcd6dd-61fc-4b15-8c34-b7b27775f1cd-kube-api-access-d5bk4\") pod \"nmstate-console-plugin-7fbb5f6569-8vs59\" (UID: \"61dcd6dd-61fc-4b15-8c34-b7b27775f1cd\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8vs59" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.212993 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/61dcd6dd-61fc-4b15-8c34-b7b27775f1cd-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-8vs59\" (UID: \"61dcd6dd-61fc-4b15-8c34-b7b27775f1cd\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8vs59" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.213047 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/61dcd6dd-61fc-4b15-8c34-b7b27775f1cd-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-8vs59\" (UID: \"61dcd6dd-61fc-4b15-8c34-b7b27775f1cd\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8vs59" Dec 09 15:43:15 crc kubenswrapper[4894]: E1209 15:43:15.213200 4894 secret.go:188] Couldn't get secret openshift-nmstate/plugin-serving-cert: secret "plugin-serving-cert" not found Dec 09 15:43:15 crc kubenswrapper[4894]: E1209 15:43:15.213263 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/61dcd6dd-61fc-4b15-8c34-b7b27775f1cd-plugin-serving-cert podName:61dcd6dd-61fc-4b15-8c34-b7b27775f1cd nodeName:}" failed. No retries permitted until 2025-12-09 15:43:15.713244038 +0000 UTC m=+690.032454707 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "plugin-serving-cert" (UniqueName: "kubernetes.io/secret/61dcd6dd-61fc-4b15-8c34-b7b27775f1cd-plugin-serving-cert") pod "nmstate-console-plugin-7fbb5f6569-8vs59" (UID: "61dcd6dd-61fc-4b15-8c34-b7b27775f1cd") : secret "plugin-serving-cert" not found Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.214132 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/61dcd6dd-61fc-4b15-8c34-b7b27775f1cd-nginx-conf\") pod \"nmstate-console-plugin-7fbb5f6569-8vs59\" (UID: \"61dcd6dd-61fc-4b15-8c34-b7b27775f1cd\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8vs59" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.221572 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-4vf4p" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.245401 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5bk4\" (UniqueName: \"kubernetes.io/projected/61dcd6dd-61fc-4b15-8c34-b7b27775f1cd-kube-api-access-d5bk4\") pod \"nmstate-console-plugin-7fbb5f6569-8vs59\" (UID: \"61dcd6dd-61fc-4b15-8c34-b7b27775f1cd\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8vs59" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.252864 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-67b4bcfb7d-bzqqn"] Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.253692 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: W1209 15:43:15.258869 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod34e69c07_d85c_4a34_afe7_287647f01876.slice/crio-fa80a2f056bbc4cef7fe03b494d4c1bdab6721c83ed4a7b402b8ccc7b2720510 WatchSource:0}: Error finding container fa80a2f056bbc4cef7fe03b494d4c1bdab6721c83ed4a7b402b8ccc7b2720510: Status 404 returned error can't find the container with id fa80a2f056bbc4cef7fe03b494d4c1bdab6721c83ed4a7b402b8ccc7b2720510 Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.264804 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-67b4bcfb7d-bzqqn"] Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.413096 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-7f946cbc9-tclzs"] Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.415279 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b3ce8b73-3d5e-4161-9dfa-140606c1e00d-trusted-ca-bundle\") pod \"console-67b4bcfb7d-bzqqn\" (UID: \"b3ce8b73-3d5e-4161-9dfa-140606c1e00d\") " pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.415341 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hfctt\" (UniqueName: \"kubernetes.io/projected/b3ce8b73-3d5e-4161-9dfa-140606c1e00d-kube-api-access-hfctt\") pod \"console-67b4bcfb7d-bzqqn\" (UID: \"b3ce8b73-3d5e-4161-9dfa-140606c1e00d\") " pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.415374 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b3ce8b73-3d5e-4161-9dfa-140606c1e00d-console-config\") pod \"console-67b4bcfb7d-bzqqn\" (UID: \"b3ce8b73-3d5e-4161-9dfa-140606c1e00d\") " pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.415398 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b3ce8b73-3d5e-4161-9dfa-140606c1e00d-oauth-serving-cert\") pod \"console-67b4bcfb7d-bzqqn\" (UID: \"b3ce8b73-3d5e-4161-9dfa-140606c1e00d\") " pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.415448 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b3ce8b73-3d5e-4161-9dfa-140606c1e00d-console-serving-cert\") pod \"console-67b4bcfb7d-bzqqn\" (UID: \"b3ce8b73-3d5e-4161-9dfa-140606c1e00d\") " pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.415480 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b3ce8b73-3d5e-4161-9dfa-140606c1e00d-service-ca\") pod \"console-67b4bcfb7d-bzqqn\" (UID: \"b3ce8b73-3d5e-4161-9dfa-140606c1e00d\") " pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.415510 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b3ce8b73-3d5e-4161-9dfa-140606c1e00d-console-oauth-config\") pod \"console-67b4bcfb7d-bzqqn\" (UID: \"b3ce8b73-3d5e-4161-9dfa-140606c1e00d\") " pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: W1209 15:43:15.418137 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81fe83b5_37f9_499e_af80_602c302d905c.slice/crio-daecc62e3e8f35d41065e27be0574b2c550ec52d1d9340ed991dbd41e610436b WatchSource:0}: Error finding container daecc62e3e8f35d41065e27be0574b2c550ec52d1d9340ed991dbd41e610436b: Status 404 returned error can't find the container with id daecc62e3e8f35d41065e27be0574b2c550ec52d1d9340ed991dbd41e610436b Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.517315 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b3ce8b73-3d5e-4161-9dfa-140606c1e00d-console-oauth-config\") pod \"console-67b4bcfb7d-bzqqn\" (UID: \"b3ce8b73-3d5e-4161-9dfa-140606c1e00d\") " pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.517383 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b3ce8b73-3d5e-4161-9dfa-140606c1e00d-trusted-ca-bundle\") pod \"console-67b4bcfb7d-bzqqn\" (UID: \"b3ce8b73-3d5e-4161-9dfa-140606c1e00d\") " pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.517430 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hfctt\" (UniqueName: \"kubernetes.io/projected/b3ce8b73-3d5e-4161-9dfa-140606c1e00d-kube-api-access-hfctt\") pod \"console-67b4bcfb7d-bzqqn\" (UID: \"b3ce8b73-3d5e-4161-9dfa-140606c1e00d\") " pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.517457 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b3ce8b73-3d5e-4161-9dfa-140606c1e00d-console-config\") pod \"console-67b4bcfb7d-bzqqn\" (UID: \"b3ce8b73-3d5e-4161-9dfa-140606c1e00d\") " pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.517483 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b3ce8b73-3d5e-4161-9dfa-140606c1e00d-oauth-serving-cert\") pod \"console-67b4bcfb7d-bzqqn\" (UID: \"b3ce8b73-3d5e-4161-9dfa-140606c1e00d\") " pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.517523 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b3ce8b73-3d5e-4161-9dfa-140606c1e00d-console-serving-cert\") pod \"console-67b4bcfb7d-bzqqn\" (UID: \"b3ce8b73-3d5e-4161-9dfa-140606c1e00d\") " pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.517555 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b3ce8b73-3d5e-4161-9dfa-140606c1e00d-service-ca\") pod \"console-67b4bcfb7d-bzqqn\" (UID: \"b3ce8b73-3d5e-4161-9dfa-140606c1e00d\") " pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.517574 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/2cbfef40-9c05-4039-b1db-d47e119ab8da-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-5zmgn\" (UID: \"2cbfef40-9c05-4039-b1db-d47e119ab8da\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5zmgn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.518442 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/b3ce8b73-3d5e-4161-9dfa-140606c1e00d-console-config\") pod \"console-67b4bcfb7d-bzqqn\" (UID: \"b3ce8b73-3d5e-4161-9dfa-140606c1e00d\") " pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.518472 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/b3ce8b73-3d5e-4161-9dfa-140606c1e00d-oauth-serving-cert\") pod \"console-67b4bcfb7d-bzqqn\" (UID: \"b3ce8b73-3d5e-4161-9dfa-140606c1e00d\") " pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.518568 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/b3ce8b73-3d5e-4161-9dfa-140606c1e00d-trusted-ca-bundle\") pod \"console-67b4bcfb7d-bzqqn\" (UID: \"b3ce8b73-3d5e-4161-9dfa-140606c1e00d\") " pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.519778 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b3ce8b73-3d5e-4161-9dfa-140606c1e00d-service-ca\") pod \"console-67b4bcfb7d-bzqqn\" (UID: \"b3ce8b73-3d5e-4161-9dfa-140606c1e00d\") " pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.521357 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/2cbfef40-9c05-4039-b1db-d47e119ab8da-tls-key-pair\") pod \"nmstate-webhook-5f6d4c5ccb-5zmgn\" (UID: \"2cbfef40-9c05-4039-b1db-d47e119ab8da\") " pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5zmgn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.522459 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/b3ce8b73-3d5e-4161-9dfa-140606c1e00d-console-serving-cert\") pod \"console-67b4bcfb7d-bzqqn\" (UID: \"b3ce8b73-3d5e-4161-9dfa-140606c1e00d\") " pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.522762 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/b3ce8b73-3d5e-4161-9dfa-140606c1e00d-console-oauth-config\") pod \"console-67b4bcfb7d-bzqqn\" (UID: \"b3ce8b73-3d5e-4161-9dfa-140606c1e00d\") " pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.534754 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hfctt\" (UniqueName: \"kubernetes.io/projected/b3ce8b73-3d5e-4161-9dfa-140606c1e00d-kube-api-access-hfctt\") pod \"console-67b4bcfb7d-bzqqn\" (UID: \"b3ce8b73-3d5e-4161-9dfa-140606c1e00d\") " pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.577115 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.720537 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/61dcd6dd-61fc-4b15-8c34-b7b27775f1cd-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-8vs59\" (UID: \"61dcd6dd-61fc-4b15-8c34-b7b27775f1cd\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8vs59" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.724406 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/61dcd6dd-61fc-4b15-8c34-b7b27775f1cd-plugin-serving-cert\") pod \"nmstate-console-plugin-7fbb5f6569-8vs59\" (UID: \"61dcd6dd-61fc-4b15-8c34-b7b27775f1cd\") " pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8vs59" Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.764585 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-67b4bcfb7d-bzqqn"] Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.768018 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5zmgn" Dec 09 15:43:15 crc kubenswrapper[4894]: W1209 15:43:15.770405 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb3ce8b73_3d5e_4161_9dfa_140606c1e00d.slice/crio-5d596951fec9d608e845b0e48831cb02ec8b1f0ecf2e400b937015a7abaa726e WatchSource:0}: Error finding container 5d596951fec9d608e845b0e48831cb02ec8b1f0ecf2e400b937015a7abaa726e: Status 404 returned error can't find the container with id 5d596951fec9d608e845b0e48831cb02ec8b1f0ecf2e400b937015a7abaa726e Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.851818 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-67b4bcfb7d-bzqqn" event={"ID":"b3ce8b73-3d5e-4161-9dfa-140606c1e00d","Type":"ContainerStarted","Data":"5d596951fec9d608e845b0e48831cb02ec8b1f0ecf2e400b937015a7abaa726e"} Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.853370 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-4vf4p" event={"ID":"34e69c07-d85c-4a34-afe7-287647f01876","Type":"ContainerStarted","Data":"fa80a2f056bbc4cef7fe03b494d4c1bdab6721c83ed4a7b402b8ccc7b2720510"} Dec 09 15:43:15 crc kubenswrapper[4894]: I1209 15:43:15.854421 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-tclzs" event={"ID":"81fe83b5-37f9-499e-af80-602c302d905c","Type":"ContainerStarted","Data":"daecc62e3e8f35d41065e27be0574b2c550ec52d1d9340ed991dbd41e610436b"} Dec 09 15:43:16 crc kubenswrapper[4894]: I1209 15:43:16.005508 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8vs59" Dec 09 15:43:16 crc kubenswrapper[4894]: I1209 15:43:16.027173 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5zmgn"] Dec 09 15:43:16 crc kubenswrapper[4894]: W1209 15:43:16.038971 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2cbfef40_9c05_4039_b1db_d47e119ab8da.slice/crio-f8dda5e4b5372af645cd01158d202d68f0557e388b56e7686e8e2aae0eb78121 WatchSource:0}: Error finding container f8dda5e4b5372af645cd01158d202d68f0557e388b56e7686e8e2aae0eb78121: Status 404 returned error can't find the container with id f8dda5e4b5372af645cd01158d202d68f0557e388b56e7686e8e2aae0eb78121 Dec 09 15:43:16 crc kubenswrapper[4894]: I1209 15:43:16.438697 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8vs59"] Dec 09 15:43:16 crc kubenswrapper[4894]: W1209 15:43:16.451823 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod61dcd6dd_61fc_4b15_8c34_b7b27775f1cd.slice/crio-b37acd1365f114df46da50d02eefc4b0cd404dfa7b68f1b3d225ae15e1b5ade6 WatchSource:0}: Error finding container b37acd1365f114df46da50d02eefc4b0cd404dfa7b68f1b3d225ae15e1b5ade6: Status 404 returned error can't find the container with id b37acd1365f114df46da50d02eefc4b0cd404dfa7b68f1b3d225ae15e1b5ade6 Dec 09 15:43:16 crc kubenswrapper[4894]: I1209 15:43:16.859832 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8vs59" event={"ID":"61dcd6dd-61fc-4b15-8c34-b7b27775f1cd","Type":"ContainerStarted","Data":"b37acd1365f114df46da50d02eefc4b0cd404dfa7b68f1b3d225ae15e1b5ade6"} Dec 09 15:43:16 crc kubenswrapper[4894]: I1209 15:43:16.862209 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-67b4bcfb7d-bzqqn" event={"ID":"b3ce8b73-3d5e-4161-9dfa-140606c1e00d","Type":"ContainerStarted","Data":"6c565252b77d5a098ed422389e8432484a1f1884115871472958795269df90b5"} Dec 09 15:43:16 crc kubenswrapper[4894]: I1209 15:43:16.863159 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5zmgn" event={"ID":"2cbfef40-9c05-4039-b1db-d47e119ab8da","Type":"ContainerStarted","Data":"f8dda5e4b5372af645cd01158d202d68f0557e388b56e7686e8e2aae0eb78121"} Dec 09 15:43:16 crc kubenswrapper[4894]: I1209 15:43:16.886028 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-67b4bcfb7d-bzqqn" podStartSLOduration=1.886010586 podStartE2EDuration="1.886010586s" podCreationTimestamp="2025-12-09 15:43:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:43:16.879789978 +0000 UTC m=+691.199000647" watchObservedRunningTime="2025-12-09 15:43:16.886010586 +0000 UTC m=+691.205221255" Dec 09 15:43:17 crc kubenswrapper[4894]: I1209 15:43:17.869601 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5zmgn" event={"ID":"2cbfef40-9c05-4039-b1db-d47e119ab8da","Type":"ContainerStarted","Data":"2f25fa4797c0a8a6c62d336a299006d3753b11d9a29fe920fd764610768b273d"} Dec 09 15:43:17 crc kubenswrapper[4894]: I1209 15:43:17.869871 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5zmgn" Dec 09 15:43:17 crc kubenswrapper[4894]: I1209 15:43:17.872749 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-4vf4p" event={"ID":"34e69c07-d85c-4a34-afe7-287647f01876","Type":"ContainerStarted","Data":"1144891dac430f130f211a1e1a6299c3f5c0c53a48317a01157ecd3823c9e392"} Dec 09 15:43:17 crc kubenswrapper[4894]: I1209 15:43:17.887427 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5zmgn" podStartSLOduration=2.3534509679999998 podStartE2EDuration="3.887397248s" podCreationTimestamp="2025-12-09 15:43:14 +0000 UTC" firstStartedPulling="2025-12-09 15:43:16.041424564 +0000 UTC m=+690.360635233" lastFinishedPulling="2025-12-09 15:43:17.575370844 +0000 UTC m=+691.894581513" observedRunningTime="2025-12-09 15:43:17.887245904 +0000 UTC m=+692.206456593" watchObservedRunningTime="2025-12-09 15:43:17.887397248 +0000 UTC m=+692.206607917" Dec 09 15:43:18 crc kubenswrapper[4894]: I1209 15:43:18.879860 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8vs59" event={"ID":"61dcd6dd-61fc-4b15-8c34-b7b27775f1cd","Type":"ContainerStarted","Data":"9d1f945a10d6c9c2b309a30ebf0ebe06b2f727ecf2d4754577d9eb18cd828e90"} Dec 09 15:43:18 crc kubenswrapper[4894]: I1209 15:43:18.880460 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-4vf4p" Dec 09 15:43:18 crc kubenswrapper[4894]: I1209 15:43:18.894867 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-4vf4p" podStartSLOduration=2.563894488 podStartE2EDuration="4.894848636s" podCreationTimestamp="2025-12-09 15:43:14 +0000 UTC" firstStartedPulling="2025-12-09 15:43:15.265254503 +0000 UTC m=+689.584465172" lastFinishedPulling="2025-12-09 15:43:17.596208651 +0000 UTC m=+691.915419320" observedRunningTime="2025-12-09 15:43:17.907722412 +0000 UTC m=+692.226933111" watchObservedRunningTime="2025-12-09 15:43:18.894848636 +0000 UTC m=+693.214059305" Dec 09 15:43:18 crc kubenswrapper[4894]: I1209 15:43:18.896922 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-7fbb5f6569-8vs59" podStartSLOduration=1.789481938 podStartE2EDuration="3.896912431s" podCreationTimestamp="2025-12-09 15:43:15 +0000 UTC" firstStartedPulling="2025-12-09 15:43:16.457221753 +0000 UTC m=+690.776432422" lastFinishedPulling="2025-12-09 15:43:18.564652246 +0000 UTC m=+692.883862915" observedRunningTime="2025-12-09 15:43:18.89352909 +0000 UTC m=+693.212739769" watchObservedRunningTime="2025-12-09 15:43:18.896912431 +0000 UTC m=+693.216123100" Dec 09 15:43:22 crc kubenswrapper[4894]: I1209 15:43:22.914110 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-tclzs" event={"ID":"81fe83b5-37f9-499e-af80-602c302d905c","Type":"ContainerStarted","Data":"030e3e54e537b12b8eda0a16bd59fedc278e1d80518f61dda687a40d6cfcfa5a"} Dec 09 15:43:25 crc kubenswrapper[4894]: I1209 15:43:25.248366 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-4vf4p" Dec 09 15:43:25 crc kubenswrapper[4894]: I1209 15:43:25.577891 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:25 crc kubenswrapper[4894]: I1209 15:43:25.577979 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:25 crc kubenswrapper[4894]: I1209 15:43:25.584770 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:25 crc kubenswrapper[4894]: I1209 15:43:25.937365 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-tclzs" event={"ID":"81fe83b5-37f9-499e-af80-602c302d905c","Type":"ContainerStarted","Data":"bee60acab00167db2a66b249a59832d93fd79e78a2317820d0f620b5d116c563"} Dec 09 15:43:25 crc kubenswrapper[4894]: I1209 15:43:25.943544 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-67b4bcfb7d-bzqqn" Dec 09 15:43:25 crc kubenswrapper[4894]: I1209 15:43:25.969459 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-7f946cbc9-tclzs" podStartSLOduration=2.530617803 podStartE2EDuration="11.969440817s" podCreationTimestamp="2025-12-09 15:43:14 +0000 UTC" firstStartedPulling="2025-12-09 15:43:15.42050463 +0000 UTC m=+689.739715299" lastFinishedPulling="2025-12-09 15:43:24.859327604 +0000 UTC m=+699.178538313" observedRunningTime="2025-12-09 15:43:25.964286486 +0000 UTC m=+700.283497165" watchObservedRunningTime="2025-12-09 15:43:25.969440817 +0000 UTC m=+700.288651486" Dec 09 15:43:26 crc kubenswrapper[4894]: I1209 15:43:26.011266 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-crhvg"] Dec 09 15:43:35 crc kubenswrapper[4894]: I1209 15:43:35.777159 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-5f6d4c5ccb-5zmgn" Dec 09 15:43:42 crc kubenswrapper[4894]: I1209 15:43:42.257747 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:43:42 crc kubenswrapper[4894]: I1209 15:43:42.258058 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:43:50 crc kubenswrapper[4894]: I1209 15:43:50.552749 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl"] Dec 09 15:43:50 crc kubenswrapper[4894]: I1209 15:43:50.554735 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl" Dec 09 15:43:50 crc kubenswrapper[4894]: I1209 15:43:50.557946 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Dec 09 15:43:50 crc kubenswrapper[4894]: I1209 15:43:50.564938 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl"] Dec 09 15:43:50 crc kubenswrapper[4894]: I1209 15:43:50.567968 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/423c01ef-021c-44a3-a606-df1bfef275df-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl\" (UID: \"423c01ef-021c-44a3-a606-df1bfef275df\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl" Dec 09 15:43:50 crc kubenswrapper[4894]: I1209 15:43:50.568058 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/423c01ef-021c-44a3-a606-df1bfef275df-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl\" (UID: \"423c01ef-021c-44a3-a606-df1bfef275df\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl" Dec 09 15:43:50 crc kubenswrapper[4894]: I1209 15:43:50.568109 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4c8tp\" (UniqueName: \"kubernetes.io/projected/423c01ef-021c-44a3-a606-df1bfef275df-kube-api-access-4c8tp\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl\" (UID: \"423c01ef-021c-44a3-a606-df1bfef275df\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl" Dec 09 15:43:50 crc kubenswrapper[4894]: I1209 15:43:50.668659 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/423c01ef-021c-44a3-a606-df1bfef275df-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl\" (UID: \"423c01ef-021c-44a3-a606-df1bfef275df\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl" Dec 09 15:43:50 crc kubenswrapper[4894]: I1209 15:43:50.668715 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4c8tp\" (UniqueName: \"kubernetes.io/projected/423c01ef-021c-44a3-a606-df1bfef275df-kube-api-access-4c8tp\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl\" (UID: \"423c01ef-021c-44a3-a606-df1bfef275df\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl" Dec 09 15:43:50 crc kubenswrapper[4894]: I1209 15:43:50.668755 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/423c01ef-021c-44a3-a606-df1bfef275df-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl\" (UID: \"423c01ef-021c-44a3-a606-df1bfef275df\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl" Dec 09 15:43:50 crc kubenswrapper[4894]: I1209 15:43:50.669159 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/423c01ef-021c-44a3-a606-df1bfef275df-util\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl\" (UID: \"423c01ef-021c-44a3-a606-df1bfef275df\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl" Dec 09 15:43:50 crc kubenswrapper[4894]: I1209 15:43:50.669177 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/423c01ef-021c-44a3-a606-df1bfef275df-bundle\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl\" (UID: \"423c01ef-021c-44a3-a606-df1bfef275df\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl" Dec 09 15:43:50 crc kubenswrapper[4894]: I1209 15:43:50.686280 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4c8tp\" (UniqueName: \"kubernetes.io/projected/423c01ef-021c-44a3-a606-df1bfef275df-kube-api-access-4c8tp\") pod \"af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl\" (UID: \"423c01ef-021c-44a3-a606-df1bfef275df\") " pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl" Dec 09 15:43:50 crc kubenswrapper[4894]: I1209 15:43:50.871521 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl" Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.071085 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-crhvg" podUID="9d548d70-c4af-4333-8cea-f6bef5722ce5" containerName="console" containerID="cri-o://934a294305d34c4714588f014418decc38f192e7a3074c26506ebede4edeb817" gracePeriod=15 Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.269309 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl"] Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.386837 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-crhvg_9d548d70-c4af-4333-8cea-f6bef5722ce5/console/0.log" Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.387263 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.578678 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-service-ca\") pod \"9d548d70-c4af-4333-8cea-f6bef5722ce5\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.578743 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d548d70-c4af-4333-8cea-f6bef5722ce5-console-oauth-config\") pod \"9d548d70-c4af-4333-8cea-f6bef5722ce5\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.578800 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d548d70-c4af-4333-8cea-f6bef5722ce5-console-serving-cert\") pod \"9d548d70-c4af-4333-8cea-f6bef5722ce5\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.578822 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kzmm4\" (UniqueName: \"kubernetes.io/projected/9d548d70-c4af-4333-8cea-f6bef5722ce5-kube-api-access-kzmm4\") pod \"9d548d70-c4af-4333-8cea-f6bef5722ce5\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.578872 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-trusted-ca-bundle\") pod \"9d548d70-c4af-4333-8cea-f6bef5722ce5\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.579613 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "9d548d70-c4af-4333-8cea-f6bef5722ce5" (UID: "9d548d70-c4af-4333-8cea-f6bef5722ce5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.579674 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-console-config\") pod \"9d548d70-c4af-4333-8cea-f6bef5722ce5\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.579729 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-oauth-serving-cert\") pod \"9d548d70-c4af-4333-8cea-f6bef5722ce5\" (UID: \"9d548d70-c4af-4333-8cea-f6bef5722ce5\") " Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.579985 4894 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.580117 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-console-config" (OuterVolumeSpecName: "console-config") pod "9d548d70-c4af-4333-8cea-f6bef5722ce5" (UID: "9d548d70-c4af-4333-8cea-f6bef5722ce5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.580479 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-service-ca" (OuterVolumeSpecName: "service-ca") pod "9d548d70-c4af-4333-8cea-f6bef5722ce5" (UID: "9d548d70-c4af-4333-8cea-f6bef5722ce5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.580500 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "9d548d70-c4af-4333-8cea-f6bef5722ce5" (UID: "9d548d70-c4af-4333-8cea-f6bef5722ce5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.583538 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d548d70-c4af-4333-8cea-f6bef5722ce5-kube-api-access-kzmm4" (OuterVolumeSpecName: "kube-api-access-kzmm4") pod "9d548d70-c4af-4333-8cea-f6bef5722ce5" (UID: "9d548d70-c4af-4333-8cea-f6bef5722ce5"). InnerVolumeSpecName "kube-api-access-kzmm4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.583743 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d548d70-c4af-4333-8cea-f6bef5722ce5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "9d548d70-c4af-4333-8cea-f6bef5722ce5" (UID: "9d548d70-c4af-4333-8cea-f6bef5722ce5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.583919 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d548d70-c4af-4333-8cea-f6bef5722ce5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "9d548d70-c4af-4333-8cea-f6bef5722ce5" (UID: "9d548d70-c4af-4333-8cea-f6bef5722ce5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.680831 4894 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.680880 4894 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-service-ca\") on node \"crc\" DevicePath \"\"" Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.680889 4894 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/9d548d70-c4af-4333-8cea-f6bef5722ce5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.680899 4894 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/9d548d70-c4af-4333-8cea-f6bef5722ce5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.680907 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kzmm4\" (UniqueName: \"kubernetes.io/projected/9d548d70-c4af-4333-8cea-f6bef5722ce5-kube-api-access-kzmm4\") on node \"crc\" DevicePath \"\"" Dec 09 15:43:51 crc kubenswrapper[4894]: I1209 15:43:51.680915 4894 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/9d548d70-c4af-4333-8cea-f6bef5722ce5-console-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:43:52 crc kubenswrapper[4894]: I1209 15:43:52.079389 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-crhvg_9d548d70-c4af-4333-8cea-f6bef5722ce5/console/0.log" Dec 09 15:43:52 crc kubenswrapper[4894]: I1209 15:43:52.079460 4894 generic.go:334] "Generic (PLEG): container finished" podID="9d548d70-c4af-4333-8cea-f6bef5722ce5" containerID="934a294305d34c4714588f014418decc38f192e7a3074c26506ebede4edeb817" exitCode=2 Dec 09 15:43:52 crc kubenswrapper[4894]: I1209 15:43:52.079525 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-crhvg" Dec 09 15:43:52 crc kubenswrapper[4894]: I1209 15:43:52.079570 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-crhvg" event={"ID":"9d548d70-c4af-4333-8cea-f6bef5722ce5","Type":"ContainerDied","Data":"934a294305d34c4714588f014418decc38f192e7a3074c26506ebede4edeb817"} Dec 09 15:43:52 crc kubenswrapper[4894]: I1209 15:43:52.079596 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-crhvg" event={"ID":"9d548d70-c4af-4333-8cea-f6bef5722ce5","Type":"ContainerDied","Data":"3a9d5ae85cba8e07af05cd9ed81836acb2a42edd0fafb15888d327bd0fe8f5f5"} Dec 09 15:43:52 crc kubenswrapper[4894]: I1209 15:43:52.079616 4894 scope.go:117] "RemoveContainer" containerID="934a294305d34c4714588f014418decc38f192e7a3074c26506ebede4edeb817" Dec 09 15:43:52 crc kubenswrapper[4894]: I1209 15:43:52.081435 4894 generic.go:334] "Generic (PLEG): container finished" podID="423c01ef-021c-44a3-a606-df1bfef275df" containerID="7144fc5268295ab5ced1855c7c7a81cd3fed7850d361cb3d1be0c22069d12b63" exitCode=0 Dec 09 15:43:52 crc kubenswrapper[4894]: I1209 15:43:52.081458 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl" event={"ID":"423c01ef-021c-44a3-a606-df1bfef275df","Type":"ContainerDied","Data":"7144fc5268295ab5ced1855c7c7a81cd3fed7850d361cb3d1be0c22069d12b63"} Dec 09 15:43:52 crc kubenswrapper[4894]: I1209 15:43:52.081472 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl" event={"ID":"423c01ef-021c-44a3-a606-df1bfef275df","Type":"ContainerStarted","Data":"e5f82453c88300ac47072a6e37d2aa5d7afef8fd3e97618a1cfc3beb74fb02f1"} Dec 09 15:43:52 crc kubenswrapper[4894]: I1209 15:43:52.097970 4894 scope.go:117] "RemoveContainer" containerID="934a294305d34c4714588f014418decc38f192e7a3074c26506ebede4edeb817" Dec 09 15:43:52 crc kubenswrapper[4894]: E1209 15:43:52.098477 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"934a294305d34c4714588f014418decc38f192e7a3074c26506ebede4edeb817\": container with ID starting with 934a294305d34c4714588f014418decc38f192e7a3074c26506ebede4edeb817 not found: ID does not exist" containerID="934a294305d34c4714588f014418decc38f192e7a3074c26506ebede4edeb817" Dec 09 15:43:52 crc kubenswrapper[4894]: I1209 15:43:52.098532 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"934a294305d34c4714588f014418decc38f192e7a3074c26506ebede4edeb817"} err="failed to get container status \"934a294305d34c4714588f014418decc38f192e7a3074c26506ebede4edeb817\": rpc error: code = NotFound desc = could not find container \"934a294305d34c4714588f014418decc38f192e7a3074c26506ebede4edeb817\": container with ID starting with 934a294305d34c4714588f014418decc38f192e7a3074c26506ebede4edeb817 not found: ID does not exist" Dec 09 15:43:52 crc kubenswrapper[4894]: I1209 15:43:52.119867 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-crhvg"] Dec 09 15:43:52 crc kubenswrapper[4894]: I1209 15:43:52.123314 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-crhvg"] Dec 09 15:43:54 crc kubenswrapper[4894]: I1209 15:43:54.107107 4894 generic.go:334] "Generic (PLEG): container finished" podID="423c01ef-021c-44a3-a606-df1bfef275df" containerID="ba507c43ade390a6504bca0d618c0a39f01291aa4dea7b09d5d0460c533ba1dd" exitCode=0 Dec 09 15:43:54 crc kubenswrapper[4894]: I1209 15:43:54.119755 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d548d70-c4af-4333-8cea-f6bef5722ce5" path="/var/lib/kubelet/pods/9d548d70-c4af-4333-8cea-f6bef5722ce5/volumes" Dec 09 15:43:54 crc kubenswrapper[4894]: I1209 15:43:54.120333 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl" event={"ID":"423c01ef-021c-44a3-a606-df1bfef275df","Type":"ContainerDied","Data":"ba507c43ade390a6504bca0d618c0a39f01291aa4dea7b09d5d0460c533ba1dd"} Dec 09 15:43:55 crc kubenswrapper[4894]: I1209 15:43:55.115548 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl" event={"ID":"423c01ef-021c-44a3-a606-df1bfef275df","Type":"ContainerStarted","Data":"f578e1299fe0c639b2bc3dc3785cadc06a3ee24bbdcc43862305a19e10e4f091"} Dec 09 15:43:55 crc kubenswrapper[4894]: I1209 15:43:55.137019 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl" podStartSLOduration=4.103952088 podStartE2EDuration="5.136996052s" podCreationTimestamp="2025-12-09 15:43:50 +0000 UTC" firstStartedPulling="2025-12-09 15:43:52.08298351 +0000 UTC m=+726.402194179" lastFinishedPulling="2025-12-09 15:43:53.116027474 +0000 UTC m=+727.435238143" observedRunningTime="2025-12-09 15:43:55.132193384 +0000 UTC m=+729.451404053" watchObservedRunningTime="2025-12-09 15:43:55.136996052 +0000 UTC m=+729.456206721" Dec 09 15:43:56 crc kubenswrapper[4894]: I1209 15:43:56.122070 4894 generic.go:334] "Generic (PLEG): container finished" podID="423c01ef-021c-44a3-a606-df1bfef275df" containerID="f578e1299fe0c639b2bc3dc3785cadc06a3ee24bbdcc43862305a19e10e4f091" exitCode=0 Dec 09 15:43:56 crc kubenswrapper[4894]: I1209 15:43:56.122128 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl" event={"ID":"423c01ef-021c-44a3-a606-df1bfef275df","Type":"ContainerDied","Data":"f578e1299fe0c639b2bc3dc3785cadc06a3ee24bbdcc43862305a19e10e4f091"} Dec 09 15:43:57 crc kubenswrapper[4894]: I1209 15:43:57.333577 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl" Dec 09 15:43:57 crc kubenswrapper[4894]: I1209 15:43:57.460095 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/423c01ef-021c-44a3-a606-df1bfef275df-bundle\") pod \"423c01ef-021c-44a3-a606-df1bfef275df\" (UID: \"423c01ef-021c-44a3-a606-df1bfef275df\") " Dec 09 15:43:57 crc kubenswrapper[4894]: I1209 15:43:57.460157 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4c8tp\" (UniqueName: \"kubernetes.io/projected/423c01ef-021c-44a3-a606-df1bfef275df-kube-api-access-4c8tp\") pod \"423c01ef-021c-44a3-a606-df1bfef275df\" (UID: \"423c01ef-021c-44a3-a606-df1bfef275df\") " Dec 09 15:43:57 crc kubenswrapper[4894]: I1209 15:43:57.460228 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/423c01ef-021c-44a3-a606-df1bfef275df-util\") pod \"423c01ef-021c-44a3-a606-df1bfef275df\" (UID: \"423c01ef-021c-44a3-a606-df1bfef275df\") " Dec 09 15:43:57 crc kubenswrapper[4894]: I1209 15:43:57.461842 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/423c01ef-021c-44a3-a606-df1bfef275df-bundle" (OuterVolumeSpecName: "bundle") pod "423c01ef-021c-44a3-a606-df1bfef275df" (UID: "423c01ef-021c-44a3-a606-df1bfef275df"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:43:57 crc kubenswrapper[4894]: I1209 15:43:57.467224 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/423c01ef-021c-44a3-a606-df1bfef275df-kube-api-access-4c8tp" (OuterVolumeSpecName: "kube-api-access-4c8tp") pod "423c01ef-021c-44a3-a606-df1bfef275df" (UID: "423c01ef-021c-44a3-a606-df1bfef275df"). InnerVolumeSpecName "kube-api-access-4c8tp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:43:57 crc kubenswrapper[4894]: I1209 15:43:57.471463 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/423c01ef-021c-44a3-a606-df1bfef275df-util" (OuterVolumeSpecName: "util") pod "423c01ef-021c-44a3-a606-df1bfef275df" (UID: "423c01ef-021c-44a3-a606-df1bfef275df"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:43:57 crc kubenswrapper[4894]: I1209 15:43:57.562065 4894 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/423c01ef-021c-44a3-a606-df1bfef275df-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:43:57 crc kubenswrapper[4894]: I1209 15:43:57.562117 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4c8tp\" (UniqueName: \"kubernetes.io/projected/423c01ef-021c-44a3-a606-df1bfef275df-kube-api-access-4c8tp\") on node \"crc\" DevicePath \"\"" Dec 09 15:43:57 crc kubenswrapper[4894]: I1209 15:43:57.562138 4894 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/423c01ef-021c-44a3-a606-df1bfef275df-util\") on node \"crc\" DevicePath \"\"" Dec 09 15:43:58 crc kubenswrapper[4894]: I1209 15:43:58.138160 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl" event={"ID":"423c01ef-021c-44a3-a606-df1bfef275df","Type":"ContainerDied","Data":"e5f82453c88300ac47072a6e37d2aa5d7afef8fd3e97618a1cfc3beb74fb02f1"} Dec 09 15:43:58 crc kubenswrapper[4894]: I1209 15:43:58.138208 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e5f82453c88300ac47072a6e37d2aa5d7afef8fd3e97618a1cfc3beb74fb02f1" Dec 09 15:43:58 crc kubenswrapper[4894]: I1209 15:43:58.138240 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl" Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.707399 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-85ff77c5c9-mzmjf"] Dec 09 15:44:06 crc kubenswrapper[4894]: E1209 15:44:06.708164 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d548d70-c4af-4333-8cea-f6bef5722ce5" containerName="console" Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.708178 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d548d70-c4af-4333-8cea-f6bef5722ce5" containerName="console" Dec 09 15:44:06 crc kubenswrapper[4894]: E1209 15:44:06.708199 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="423c01ef-021c-44a3-a606-df1bfef275df" containerName="pull" Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.708206 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="423c01ef-021c-44a3-a606-df1bfef275df" containerName="pull" Dec 09 15:44:06 crc kubenswrapper[4894]: E1209 15:44:06.708213 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="423c01ef-021c-44a3-a606-df1bfef275df" containerName="util" Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.708220 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="423c01ef-021c-44a3-a606-df1bfef275df" containerName="util" Dec 09 15:44:06 crc kubenswrapper[4894]: E1209 15:44:06.708230 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="423c01ef-021c-44a3-a606-df1bfef275df" containerName="extract" Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.708237 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="423c01ef-021c-44a3-a606-df1bfef275df" containerName="extract" Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.708366 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d548d70-c4af-4333-8cea-f6bef5722ce5" containerName="console" Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.708381 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="423c01ef-021c-44a3-a606-df1bfef275df" containerName="extract" Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.708826 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-85ff77c5c9-mzmjf" Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.710596 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.710613 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.711393 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.711433 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.713909 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-sxfzh" Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.722496 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-85ff77c5c9-mzmjf"] Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.774492 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2d23be67-e833-46bd-a293-a41fc8144ca1-webhook-cert\") pod \"metallb-operator-controller-manager-85ff77c5c9-mzmjf\" (UID: \"2d23be67-e833-46bd-a293-a41fc8144ca1\") " pod="metallb-system/metallb-operator-controller-manager-85ff77c5c9-mzmjf" Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.774774 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2d23be67-e833-46bd-a293-a41fc8144ca1-apiservice-cert\") pod \"metallb-operator-controller-manager-85ff77c5c9-mzmjf\" (UID: \"2d23be67-e833-46bd-a293-a41fc8144ca1\") " pod="metallb-system/metallb-operator-controller-manager-85ff77c5c9-mzmjf" Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.774883 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbknp\" (UniqueName: \"kubernetes.io/projected/2d23be67-e833-46bd-a293-a41fc8144ca1-kube-api-access-lbknp\") pod \"metallb-operator-controller-manager-85ff77c5c9-mzmjf\" (UID: \"2d23be67-e833-46bd-a293-a41fc8144ca1\") " pod="metallb-system/metallb-operator-controller-manager-85ff77c5c9-mzmjf" Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.876066 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2d23be67-e833-46bd-a293-a41fc8144ca1-webhook-cert\") pod \"metallb-operator-controller-manager-85ff77c5c9-mzmjf\" (UID: \"2d23be67-e833-46bd-a293-a41fc8144ca1\") " pod="metallb-system/metallb-operator-controller-manager-85ff77c5c9-mzmjf" Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.876118 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2d23be67-e833-46bd-a293-a41fc8144ca1-apiservice-cert\") pod \"metallb-operator-controller-manager-85ff77c5c9-mzmjf\" (UID: \"2d23be67-e833-46bd-a293-a41fc8144ca1\") " pod="metallb-system/metallb-operator-controller-manager-85ff77c5c9-mzmjf" Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.876165 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbknp\" (UniqueName: \"kubernetes.io/projected/2d23be67-e833-46bd-a293-a41fc8144ca1-kube-api-access-lbknp\") pod \"metallb-operator-controller-manager-85ff77c5c9-mzmjf\" (UID: \"2d23be67-e833-46bd-a293-a41fc8144ca1\") " pod="metallb-system/metallb-operator-controller-manager-85ff77c5c9-mzmjf" Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.883956 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/2d23be67-e833-46bd-a293-a41fc8144ca1-apiservice-cert\") pod \"metallb-operator-controller-manager-85ff77c5c9-mzmjf\" (UID: \"2d23be67-e833-46bd-a293-a41fc8144ca1\") " pod="metallb-system/metallb-operator-controller-manager-85ff77c5c9-mzmjf" Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.884514 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/2d23be67-e833-46bd-a293-a41fc8144ca1-webhook-cert\") pod \"metallb-operator-controller-manager-85ff77c5c9-mzmjf\" (UID: \"2d23be67-e833-46bd-a293-a41fc8144ca1\") " pod="metallb-system/metallb-operator-controller-manager-85ff77c5c9-mzmjf" Dec 09 15:44:06 crc kubenswrapper[4894]: I1209 15:44:06.904562 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbknp\" (UniqueName: \"kubernetes.io/projected/2d23be67-e833-46bd-a293-a41fc8144ca1-kube-api-access-lbknp\") pod \"metallb-operator-controller-manager-85ff77c5c9-mzmjf\" (UID: \"2d23be67-e833-46bd-a293-a41fc8144ca1\") " pod="metallb-system/metallb-operator-controller-manager-85ff77c5c9-mzmjf" Dec 09 15:44:07 crc kubenswrapper[4894]: I1209 15:44:07.024924 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-59cdbccb6-v6f5s"] Dec 09 15:44:07 crc kubenswrapper[4894]: I1209 15:44:07.025219 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-85ff77c5c9-mzmjf" Dec 09 15:44:07 crc kubenswrapper[4894]: I1209 15:44:07.026060 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-59cdbccb6-v6f5s" Dec 09 15:44:07 crc kubenswrapper[4894]: I1209 15:44:07.029630 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 09 15:44:07 crc kubenswrapper[4894]: I1209 15:44:07.030254 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Dec 09 15:44:07 crc kubenswrapper[4894]: I1209 15:44:07.032487 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-cwnxm" Dec 09 15:44:07 crc kubenswrapper[4894]: I1209 15:44:07.052494 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-59cdbccb6-v6f5s"] Dec 09 15:44:07 crc kubenswrapper[4894]: I1209 15:44:07.179098 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/587142f6-a6a8-4500-97d5-82f0d5fc04c7-apiservice-cert\") pod \"metallb-operator-webhook-server-59cdbccb6-v6f5s\" (UID: \"587142f6-a6a8-4500-97d5-82f0d5fc04c7\") " pod="metallb-system/metallb-operator-webhook-server-59cdbccb6-v6f5s" Dec 09 15:44:07 crc kubenswrapper[4894]: I1209 15:44:07.179223 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlftp\" (UniqueName: \"kubernetes.io/projected/587142f6-a6a8-4500-97d5-82f0d5fc04c7-kube-api-access-mlftp\") pod \"metallb-operator-webhook-server-59cdbccb6-v6f5s\" (UID: \"587142f6-a6a8-4500-97d5-82f0d5fc04c7\") " pod="metallb-system/metallb-operator-webhook-server-59cdbccb6-v6f5s" Dec 09 15:44:07 crc kubenswrapper[4894]: I1209 15:44:07.179256 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/587142f6-a6a8-4500-97d5-82f0d5fc04c7-webhook-cert\") pod \"metallb-operator-webhook-server-59cdbccb6-v6f5s\" (UID: \"587142f6-a6a8-4500-97d5-82f0d5fc04c7\") " pod="metallb-system/metallb-operator-webhook-server-59cdbccb6-v6f5s" Dec 09 15:44:07 crc kubenswrapper[4894]: I1209 15:44:07.274758 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-85ff77c5c9-mzmjf"] Dec 09 15:44:07 crc kubenswrapper[4894]: W1209 15:44:07.280884 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2d23be67_e833_46bd_a293_a41fc8144ca1.slice/crio-344f7a764251e4474e9be714544f0d0c1bfdcec42d53a3cf435fc63d88c960e4 WatchSource:0}: Error finding container 344f7a764251e4474e9be714544f0d0c1bfdcec42d53a3cf435fc63d88c960e4: Status 404 returned error can't find the container with id 344f7a764251e4474e9be714544f0d0c1bfdcec42d53a3cf435fc63d88c960e4 Dec 09 15:44:07 crc kubenswrapper[4894]: I1209 15:44:07.281740 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlftp\" (UniqueName: \"kubernetes.io/projected/587142f6-a6a8-4500-97d5-82f0d5fc04c7-kube-api-access-mlftp\") pod \"metallb-operator-webhook-server-59cdbccb6-v6f5s\" (UID: \"587142f6-a6a8-4500-97d5-82f0d5fc04c7\") " pod="metallb-system/metallb-operator-webhook-server-59cdbccb6-v6f5s" Dec 09 15:44:07 crc kubenswrapper[4894]: I1209 15:44:07.281766 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/587142f6-a6a8-4500-97d5-82f0d5fc04c7-webhook-cert\") pod \"metallb-operator-webhook-server-59cdbccb6-v6f5s\" (UID: \"587142f6-a6a8-4500-97d5-82f0d5fc04c7\") " pod="metallb-system/metallb-operator-webhook-server-59cdbccb6-v6f5s" Dec 09 15:44:07 crc kubenswrapper[4894]: I1209 15:44:07.281836 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/587142f6-a6a8-4500-97d5-82f0d5fc04c7-apiservice-cert\") pod \"metallb-operator-webhook-server-59cdbccb6-v6f5s\" (UID: \"587142f6-a6a8-4500-97d5-82f0d5fc04c7\") " pod="metallb-system/metallb-operator-webhook-server-59cdbccb6-v6f5s" Dec 09 15:44:07 crc kubenswrapper[4894]: I1209 15:44:07.287276 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/587142f6-a6a8-4500-97d5-82f0d5fc04c7-apiservice-cert\") pod \"metallb-operator-webhook-server-59cdbccb6-v6f5s\" (UID: \"587142f6-a6a8-4500-97d5-82f0d5fc04c7\") " pod="metallb-system/metallb-operator-webhook-server-59cdbccb6-v6f5s" Dec 09 15:44:07 crc kubenswrapper[4894]: I1209 15:44:07.287933 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/587142f6-a6a8-4500-97d5-82f0d5fc04c7-webhook-cert\") pod \"metallb-operator-webhook-server-59cdbccb6-v6f5s\" (UID: \"587142f6-a6a8-4500-97d5-82f0d5fc04c7\") " pod="metallb-system/metallb-operator-webhook-server-59cdbccb6-v6f5s" Dec 09 15:44:07 crc kubenswrapper[4894]: I1209 15:44:07.302523 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlftp\" (UniqueName: \"kubernetes.io/projected/587142f6-a6a8-4500-97d5-82f0d5fc04c7-kube-api-access-mlftp\") pod \"metallb-operator-webhook-server-59cdbccb6-v6f5s\" (UID: \"587142f6-a6a8-4500-97d5-82f0d5fc04c7\") " pod="metallb-system/metallb-operator-webhook-server-59cdbccb6-v6f5s" Dec 09 15:44:07 crc kubenswrapper[4894]: I1209 15:44:07.345080 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-59cdbccb6-v6f5s" Dec 09 15:44:07 crc kubenswrapper[4894]: I1209 15:44:07.578682 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-59cdbccb6-v6f5s"] Dec 09 15:44:07 crc kubenswrapper[4894]: W1209 15:44:07.583260 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod587142f6_a6a8_4500_97d5_82f0d5fc04c7.slice/crio-6934a84c0574c673e677ab6e382d311ba3c7797e6d5a3c81f2d3596d8eecbed3 WatchSource:0}: Error finding container 6934a84c0574c673e677ab6e382d311ba3c7797e6d5a3c81f2d3596d8eecbed3: Status 404 returned error can't find the container with id 6934a84c0574c673e677ab6e382d311ba3c7797e6d5a3c81f2d3596d8eecbed3 Dec 09 15:44:08 crc kubenswrapper[4894]: I1209 15:44:08.190397 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-59cdbccb6-v6f5s" event={"ID":"587142f6-a6a8-4500-97d5-82f0d5fc04c7","Type":"ContainerStarted","Data":"6934a84c0574c673e677ab6e382d311ba3c7797e6d5a3c81f2d3596d8eecbed3"} Dec 09 15:44:08 crc kubenswrapper[4894]: I1209 15:44:08.191875 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-85ff77c5c9-mzmjf" event={"ID":"2d23be67-e833-46bd-a293-a41fc8144ca1","Type":"ContainerStarted","Data":"344f7a764251e4474e9be714544f0d0c1bfdcec42d53a3cf435fc63d88c960e4"} Dec 09 15:44:11 crc kubenswrapper[4894]: I1209 15:44:11.213009 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-85ff77c5c9-mzmjf" event={"ID":"2d23be67-e833-46bd-a293-a41fc8144ca1","Type":"ContainerStarted","Data":"2c38e5860c6e0cca71b0d130a7e2971d4012c7d8c9f86bc4f5f74affd4faff14"} Dec 09 15:44:11 crc kubenswrapper[4894]: I1209 15:44:11.213240 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-85ff77c5c9-mzmjf" Dec 09 15:44:11 crc kubenswrapper[4894]: I1209 15:44:11.239883 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-85ff77c5c9-mzmjf" podStartSLOduration=2.399030325 podStartE2EDuration="5.239868723s" podCreationTimestamp="2025-12-09 15:44:06 +0000 UTC" firstStartedPulling="2025-12-09 15:44:07.283169462 +0000 UTC m=+741.602380131" lastFinishedPulling="2025-12-09 15:44:10.12400786 +0000 UTC m=+744.443218529" observedRunningTime="2025-12-09 15:44:11.237781698 +0000 UTC m=+745.556992377" watchObservedRunningTime="2025-12-09 15:44:11.239868723 +0000 UTC m=+745.559079392" Dec 09 15:44:12 crc kubenswrapper[4894]: I1209 15:44:12.257356 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:44:12 crc kubenswrapper[4894]: I1209 15:44:12.258569 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:44:12 crc kubenswrapper[4894]: I1209 15:44:12.258732 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 15:44:12 crc kubenswrapper[4894]: I1209 15:44:12.259378 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"bf0d8a25593f3059a7a7f1286d689a889938309e9429a43b4edb9ebe0cc7eecf"} pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 15:44:12 crc kubenswrapper[4894]: I1209 15:44:12.259569 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" containerID="cri-o://bf0d8a25593f3059a7a7f1286d689a889938309e9429a43b4edb9ebe0cc7eecf" gracePeriod=600 Dec 09 15:44:13 crc kubenswrapper[4894]: I1209 15:44:13.228216 4894 generic.go:334] "Generic (PLEG): container finished" podID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerID="bf0d8a25593f3059a7a7f1286d689a889938309e9429a43b4edb9ebe0cc7eecf" exitCode=0 Dec 09 15:44:13 crc kubenswrapper[4894]: I1209 15:44:13.228786 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerDied","Data":"bf0d8a25593f3059a7a7f1286d689a889938309e9429a43b4edb9ebe0cc7eecf"} Dec 09 15:44:13 crc kubenswrapper[4894]: I1209 15:44:13.228814 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerStarted","Data":"3c92ffd29e4b95c91cb30f4577bf2eab4718185324fe428051f5af6db9d8c9a8"} Dec 09 15:44:13 crc kubenswrapper[4894]: I1209 15:44:13.228830 4894 scope.go:117] "RemoveContainer" containerID="d6ebc3b4a88a254cb43621cbd5525aefadbbb8b324c54d247db4a86dd26c746f" Dec 09 15:44:14 crc kubenswrapper[4894]: I1209 15:44:14.236733 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-59cdbccb6-v6f5s" event={"ID":"587142f6-a6a8-4500-97d5-82f0d5fc04c7","Type":"ContainerStarted","Data":"35b7ca26109b9ce47f2e36bea473984c92d3c3f63994780919346387f6037492"} Dec 09 15:44:14 crc kubenswrapper[4894]: I1209 15:44:14.237211 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-59cdbccb6-v6f5s" Dec 09 15:44:20 crc kubenswrapper[4894]: I1209 15:44:20.016544 4894 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Dec 09 15:44:27 crc kubenswrapper[4894]: I1209 15:44:27.376478 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-59cdbccb6-v6f5s" Dec 09 15:44:27 crc kubenswrapper[4894]: I1209 15:44:27.405470 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-59cdbccb6-v6f5s" podStartSLOduration=13.94972161 podStartE2EDuration="20.40545663s" podCreationTimestamp="2025-12-09 15:44:07 +0000 UTC" firstStartedPulling="2025-12-09 15:44:07.586444384 +0000 UTC m=+741.905655053" lastFinishedPulling="2025-12-09 15:44:14.042179404 +0000 UTC m=+748.361390073" observedRunningTime="2025-12-09 15:44:14.263709727 +0000 UTC m=+748.582920396" watchObservedRunningTime="2025-12-09 15:44:27.40545663 +0000 UTC m=+761.724667299" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.027604 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-85ff77c5c9-mzmjf" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.805173 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-jdblb"] Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.807550 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.813561 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.813910 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.815039 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-bzqwl" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.826301 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-5dr97"] Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.827127 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5dr97" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.829567 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.852072 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-5dr97"] Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.896773 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-8zntm"] Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.897566 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-8zntm" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.899321 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.899780 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.899985 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.900143 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-lgdv2" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.908899 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6sdl9\" (UniqueName: \"kubernetes.io/projected/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-kube-api-access-6sdl9\") pod \"frr-k8s-jdblb\" (UID: \"2b7c2f69-58e9-4aad-9900-22befc1bc6c6\") " pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.908942 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9gszz\" (UniqueName: \"kubernetes.io/projected/ba40556a-56db-491c-b0b0-e7b7f8ec1b6a-kube-api-access-9gszz\") pod \"frr-k8s-webhook-server-7fcb986d4-5dr97\" (UID: \"ba40556a-56db-491c-b0b0-e7b7f8ec1b6a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5dr97" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.909104 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ba40556a-56db-491c-b0b0-e7b7f8ec1b6a-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-5dr97\" (UID: \"ba40556a-56db-491c-b0b0-e7b7f8ec1b6a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5dr97" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.909162 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-metrics\") pod \"frr-k8s-jdblb\" (UID: \"2b7c2f69-58e9-4aad-9900-22befc1bc6c6\") " pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.909204 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-frr-startup\") pod \"frr-k8s-jdblb\" (UID: \"2b7c2f69-58e9-4aad-9900-22befc1bc6c6\") " pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.909249 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-metrics-certs\") pod \"frr-k8s-jdblb\" (UID: \"2b7c2f69-58e9-4aad-9900-22befc1bc6c6\") " pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.909293 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-frr-sockets\") pod \"frr-k8s-jdblb\" (UID: \"2b7c2f69-58e9-4aad-9900-22befc1bc6c6\") " pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.909327 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-frr-conf\") pod \"frr-k8s-jdblb\" (UID: \"2b7c2f69-58e9-4aad-9900-22befc1bc6c6\") " pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.909401 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-reloader\") pod \"frr-k8s-jdblb\" (UID: \"2b7c2f69-58e9-4aad-9900-22befc1bc6c6\") " pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.931034 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-f8648f98b-tjjzx"] Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.932014 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-tjjzx" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.934046 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Dec 09 15:44:47 crc kubenswrapper[4894]: I1209 15:44:47.942468 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-tjjzx"] Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.010059 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-metrics-certs\") pod \"frr-k8s-jdblb\" (UID: \"2b7c2f69-58e9-4aad-9900-22befc1bc6c6\") " pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.010113 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-frr-sockets\") pod \"frr-k8s-jdblb\" (UID: \"2b7c2f69-58e9-4aad-9900-22befc1bc6c6\") " pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.010127 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-frr-conf\") pod \"frr-k8s-jdblb\" (UID: \"2b7c2f69-58e9-4aad-9900-22befc1bc6c6\") " pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.010143 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-reloader\") pod \"frr-k8s-jdblb\" (UID: \"2b7c2f69-58e9-4aad-9900-22befc1bc6c6\") " pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.010167 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wp5ns\" (UniqueName: \"kubernetes.io/projected/d0f8c048-3892-4c12-a074-b7c4c5e85b87-kube-api-access-wp5ns\") pod \"speaker-8zntm\" (UID: \"d0f8c048-3892-4c12-a074-b7c4c5e85b87\") " pod="metallb-system/speaker-8zntm" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.010198 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6sdl9\" (UniqueName: \"kubernetes.io/projected/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-kube-api-access-6sdl9\") pod \"frr-k8s-jdblb\" (UID: \"2b7c2f69-58e9-4aad-9900-22befc1bc6c6\") " pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.010219 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d0f8c048-3892-4c12-a074-b7c4c5e85b87-memberlist\") pod \"speaker-8zntm\" (UID: \"d0f8c048-3892-4c12-a074-b7c4c5e85b87\") " pod="metallb-system/speaker-8zntm" Dec 09 15:44:48 crc kubenswrapper[4894]: E1209 15:44:48.010237 4894 secret.go:188] Couldn't get secret metallb-system/frr-k8s-certs-secret: secret "frr-k8s-certs-secret" not found Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.010275 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9gszz\" (UniqueName: \"kubernetes.io/projected/ba40556a-56db-491c-b0b0-e7b7f8ec1b6a-kube-api-access-9gszz\") pod \"frr-k8s-webhook-server-7fcb986d4-5dr97\" (UID: \"ba40556a-56db-491c-b0b0-e7b7f8ec1b6a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5dr97" Dec 09 15:44:48 crc kubenswrapper[4894]: E1209 15:44:48.010306 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-metrics-certs podName:2b7c2f69-58e9-4aad-9900-22befc1bc6c6 nodeName:}" failed. No retries permitted until 2025-12-09 15:44:48.510289648 +0000 UTC m=+782.829500317 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-metrics-certs") pod "frr-k8s-jdblb" (UID: "2b7c2f69-58e9-4aad-9900-22befc1bc6c6") : secret "frr-k8s-certs-secret" not found Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.010334 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d0f8c048-3892-4c12-a074-b7c4c5e85b87-metrics-certs\") pod \"speaker-8zntm\" (UID: \"d0f8c048-3892-4c12-a074-b7c4c5e85b87\") " pod="metallb-system/speaker-8zntm" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.010395 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ba40556a-56db-491c-b0b0-e7b7f8ec1b6a-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-5dr97\" (UID: \"ba40556a-56db-491c-b0b0-e7b7f8ec1b6a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5dr97" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.010436 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-metrics\") pod \"frr-k8s-jdblb\" (UID: \"2b7c2f69-58e9-4aad-9900-22befc1bc6c6\") " pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.010460 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-frr-startup\") pod \"frr-k8s-jdblb\" (UID: \"2b7c2f69-58e9-4aad-9900-22befc1bc6c6\") " pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.010476 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d0f8c048-3892-4c12-a074-b7c4c5e85b87-metallb-excludel2\") pod \"speaker-8zntm\" (UID: \"d0f8c048-3892-4c12-a074-b7c4c5e85b87\") " pod="metallb-system/speaker-8zntm" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.010623 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-frr-sockets\") pod \"frr-k8s-jdblb\" (UID: \"2b7c2f69-58e9-4aad-9900-22befc1bc6c6\") " pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.010694 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-frr-conf\") pod \"frr-k8s-jdblb\" (UID: \"2b7c2f69-58e9-4aad-9900-22befc1bc6c6\") " pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.010761 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-reloader\") pod \"frr-k8s-jdblb\" (UID: \"2b7c2f69-58e9-4aad-9900-22befc1bc6c6\") " pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:48 crc kubenswrapper[4894]: E1209 15:44:48.010807 4894 secret.go:188] Couldn't get secret metallb-system/frr-k8s-webhook-server-cert: secret "frr-k8s-webhook-server-cert" not found Dec 09 15:44:48 crc kubenswrapper[4894]: E1209 15:44:48.010830 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/ba40556a-56db-491c-b0b0-e7b7f8ec1b6a-cert podName:ba40556a-56db-491c-b0b0-e7b7f8ec1b6a nodeName:}" failed. No retries permitted until 2025-12-09 15:44:48.510823463 +0000 UTC m=+782.830034132 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/ba40556a-56db-491c-b0b0-e7b7f8ec1b6a-cert") pod "frr-k8s-webhook-server-7fcb986d4-5dr97" (UID: "ba40556a-56db-491c-b0b0-e7b7f8ec1b6a") : secret "frr-k8s-webhook-server-cert" not found Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.010942 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-metrics\") pod \"frr-k8s-jdblb\" (UID: \"2b7c2f69-58e9-4aad-9900-22befc1bc6c6\") " pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.011734 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-frr-startup\") pod \"frr-k8s-jdblb\" (UID: \"2b7c2f69-58e9-4aad-9900-22befc1bc6c6\") " pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.028230 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9gszz\" (UniqueName: \"kubernetes.io/projected/ba40556a-56db-491c-b0b0-e7b7f8ec1b6a-kube-api-access-9gszz\") pod \"frr-k8s-webhook-server-7fcb986d4-5dr97\" (UID: \"ba40556a-56db-491c-b0b0-e7b7f8ec1b6a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5dr97" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.045604 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6sdl9\" (UniqueName: \"kubernetes.io/projected/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-kube-api-access-6sdl9\") pod \"frr-k8s-jdblb\" (UID: \"2b7c2f69-58e9-4aad-9900-22befc1bc6c6\") " pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.111329 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wp5ns\" (UniqueName: \"kubernetes.io/projected/d0f8c048-3892-4c12-a074-b7c4c5e85b87-kube-api-access-wp5ns\") pod \"speaker-8zntm\" (UID: \"d0f8c048-3892-4c12-a074-b7c4c5e85b87\") " pod="metallb-system/speaker-8zntm" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.111471 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d0f8c048-3892-4c12-a074-b7c4c5e85b87-memberlist\") pod \"speaker-8zntm\" (UID: \"d0f8c048-3892-4c12-a074-b7c4c5e85b87\") " pod="metallb-system/speaker-8zntm" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.111526 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/26ecc2b2-2443-4069-bfda-934ff3835cf8-metrics-certs\") pod \"controller-f8648f98b-tjjzx\" (UID: \"26ecc2b2-2443-4069-bfda-934ff3835cf8\") " pod="metallb-system/controller-f8648f98b-tjjzx" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.111548 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tk4lh\" (UniqueName: \"kubernetes.io/projected/26ecc2b2-2443-4069-bfda-934ff3835cf8-kube-api-access-tk4lh\") pod \"controller-f8648f98b-tjjzx\" (UID: \"26ecc2b2-2443-4069-bfda-934ff3835cf8\") " pod="metallb-system/controller-f8648f98b-tjjzx" Dec 09 15:44:48 crc kubenswrapper[4894]: E1209 15:44:48.111550 4894 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 09 15:44:48 crc kubenswrapper[4894]: E1209 15:44:48.111629 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d0f8c048-3892-4c12-a074-b7c4c5e85b87-memberlist podName:d0f8c048-3892-4c12-a074-b7c4c5e85b87 nodeName:}" failed. No retries permitted until 2025-12-09 15:44:48.611614916 +0000 UTC m=+782.930825585 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d0f8c048-3892-4c12-a074-b7c4c5e85b87-memberlist") pod "speaker-8zntm" (UID: "d0f8c048-3892-4c12-a074-b7c4c5e85b87") : secret "metallb-memberlist" not found Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.111676 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d0f8c048-3892-4c12-a074-b7c4c5e85b87-metrics-certs\") pod \"speaker-8zntm\" (UID: \"d0f8c048-3892-4c12-a074-b7c4c5e85b87\") " pod="metallb-system/speaker-8zntm" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.111763 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26ecc2b2-2443-4069-bfda-934ff3835cf8-cert\") pod \"controller-f8648f98b-tjjzx\" (UID: \"26ecc2b2-2443-4069-bfda-934ff3835cf8\") " pod="metallb-system/controller-f8648f98b-tjjzx" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.111792 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d0f8c048-3892-4c12-a074-b7c4c5e85b87-metallb-excludel2\") pod \"speaker-8zntm\" (UID: \"d0f8c048-3892-4c12-a074-b7c4c5e85b87\") " pod="metallb-system/speaker-8zntm" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.112798 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/d0f8c048-3892-4c12-a074-b7c4c5e85b87-metallb-excludel2\") pod \"speaker-8zntm\" (UID: \"d0f8c048-3892-4c12-a074-b7c4c5e85b87\") " pod="metallb-system/speaker-8zntm" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.116234 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d0f8c048-3892-4c12-a074-b7c4c5e85b87-metrics-certs\") pod \"speaker-8zntm\" (UID: \"d0f8c048-3892-4c12-a074-b7c4c5e85b87\") " pod="metallb-system/speaker-8zntm" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.131091 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wp5ns\" (UniqueName: \"kubernetes.io/projected/d0f8c048-3892-4c12-a074-b7c4c5e85b87-kube-api-access-wp5ns\") pod \"speaker-8zntm\" (UID: \"d0f8c048-3892-4c12-a074-b7c4c5e85b87\") " pod="metallb-system/speaker-8zntm" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.212659 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/26ecc2b2-2443-4069-bfda-934ff3835cf8-metrics-certs\") pod \"controller-f8648f98b-tjjzx\" (UID: \"26ecc2b2-2443-4069-bfda-934ff3835cf8\") " pod="metallb-system/controller-f8648f98b-tjjzx" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.212706 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tk4lh\" (UniqueName: \"kubernetes.io/projected/26ecc2b2-2443-4069-bfda-934ff3835cf8-kube-api-access-tk4lh\") pod \"controller-f8648f98b-tjjzx\" (UID: \"26ecc2b2-2443-4069-bfda-934ff3835cf8\") " pod="metallb-system/controller-f8648f98b-tjjzx" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.212785 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26ecc2b2-2443-4069-bfda-934ff3835cf8-cert\") pod \"controller-f8648f98b-tjjzx\" (UID: \"26ecc2b2-2443-4069-bfda-934ff3835cf8\") " pod="metallb-system/controller-f8648f98b-tjjzx" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.215460 4894 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.216296 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/26ecc2b2-2443-4069-bfda-934ff3835cf8-metrics-certs\") pod \"controller-f8648f98b-tjjzx\" (UID: \"26ecc2b2-2443-4069-bfda-934ff3835cf8\") " pod="metallb-system/controller-f8648f98b-tjjzx" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.226348 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/26ecc2b2-2443-4069-bfda-934ff3835cf8-cert\") pod \"controller-f8648f98b-tjjzx\" (UID: \"26ecc2b2-2443-4069-bfda-934ff3835cf8\") " pod="metallb-system/controller-f8648f98b-tjjzx" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.228287 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tk4lh\" (UniqueName: \"kubernetes.io/projected/26ecc2b2-2443-4069-bfda-934ff3835cf8-kube-api-access-tk4lh\") pod \"controller-f8648f98b-tjjzx\" (UID: \"26ecc2b2-2443-4069-bfda-934ff3835cf8\") " pod="metallb-system/controller-f8648f98b-tjjzx" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.244624 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-f8648f98b-tjjzx" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.485928 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-f8648f98b-tjjzx"] Dec 09 15:44:48 crc kubenswrapper[4894]: W1209 15:44:48.490033 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod26ecc2b2_2443_4069_bfda_934ff3835cf8.slice/crio-a194bc4f542615a00c52817f8e6c0f31667bff2a8599795cd7ac2e6249b55ea7 WatchSource:0}: Error finding container a194bc4f542615a00c52817f8e6c0f31667bff2a8599795cd7ac2e6249b55ea7: Status 404 returned error can't find the container with id a194bc4f542615a00c52817f8e6c0f31667bff2a8599795cd7ac2e6249b55ea7 Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.515285 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ba40556a-56db-491c-b0b0-e7b7f8ec1b6a-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-5dr97\" (UID: \"ba40556a-56db-491c-b0b0-e7b7f8ec1b6a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5dr97" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.515341 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-metrics-certs\") pod \"frr-k8s-jdblb\" (UID: \"2b7c2f69-58e9-4aad-9900-22befc1bc6c6\") " pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.519698 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2b7c2f69-58e9-4aad-9900-22befc1bc6c6-metrics-certs\") pod \"frr-k8s-jdblb\" (UID: \"2b7c2f69-58e9-4aad-9900-22befc1bc6c6\") " pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.520726 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/ba40556a-56db-491c-b0b0-e7b7f8ec1b6a-cert\") pod \"frr-k8s-webhook-server-7fcb986d4-5dr97\" (UID: \"ba40556a-56db-491c-b0b0-e7b7f8ec1b6a\") " pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5dr97" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.617059 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d0f8c048-3892-4c12-a074-b7c4c5e85b87-memberlist\") pod \"speaker-8zntm\" (UID: \"d0f8c048-3892-4c12-a074-b7c4c5e85b87\") " pod="metallb-system/speaker-8zntm" Dec 09 15:44:48 crc kubenswrapper[4894]: E1209 15:44:48.617302 4894 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Dec 09 15:44:48 crc kubenswrapper[4894]: E1209 15:44:48.617402 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d0f8c048-3892-4c12-a074-b7c4c5e85b87-memberlist podName:d0f8c048-3892-4c12-a074-b7c4c5e85b87 nodeName:}" failed. No retries permitted until 2025-12-09 15:44:49.617377319 +0000 UTC m=+783.936587988 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/d0f8c048-3892-4c12-a074-b7c4c5e85b87-memberlist") pod "speaker-8zntm" (UID: "d0f8c048-3892-4c12-a074-b7c4c5e85b87") : secret "metallb-memberlist" not found Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.733519 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.740769 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5dr97" Dec 09 15:44:48 crc kubenswrapper[4894]: I1209 15:44:48.949338 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-7fcb986d4-5dr97"] Dec 09 15:44:48 crc kubenswrapper[4894]: W1209 15:44:48.954361 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podba40556a_56db_491c_b0b0_e7b7f8ec1b6a.slice/crio-1c1a8610dabde41fcae4946f4493f0e3764e225b3f5a9a04700abb97258440e5 WatchSource:0}: Error finding container 1c1a8610dabde41fcae4946f4493f0e3764e225b3f5a9a04700abb97258440e5: Status 404 returned error can't find the container with id 1c1a8610dabde41fcae4946f4493f0e3764e225b3f5a9a04700abb97258440e5 Dec 09 15:44:49 crc kubenswrapper[4894]: I1209 15:44:49.478319 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-tjjzx" event={"ID":"26ecc2b2-2443-4069-bfda-934ff3835cf8","Type":"ContainerStarted","Data":"91a38aaa2cb3c7abf82c1885a5daa198a05d3c8de2bb38126537e38593a5f01e"} Dec 09 15:44:49 crc kubenswrapper[4894]: I1209 15:44:49.478847 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-tjjzx" event={"ID":"26ecc2b2-2443-4069-bfda-934ff3835cf8","Type":"ContainerStarted","Data":"c3c93c3d94ffc967f04002988cdb64ac42bd95a061cfd1aa3482c72379ba04c6"} Dec 09 15:44:49 crc kubenswrapper[4894]: I1209 15:44:49.478865 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-f8648f98b-tjjzx" event={"ID":"26ecc2b2-2443-4069-bfda-934ff3835cf8","Type":"ContainerStarted","Data":"a194bc4f542615a00c52817f8e6c0f31667bff2a8599795cd7ac2e6249b55ea7"} Dec 09 15:44:49 crc kubenswrapper[4894]: I1209 15:44:49.478916 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-f8648f98b-tjjzx" Dec 09 15:44:49 crc kubenswrapper[4894]: I1209 15:44:49.479966 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5dr97" event={"ID":"ba40556a-56db-491c-b0b0-e7b7f8ec1b6a","Type":"ContainerStarted","Data":"1c1a8610dabde41fcae4946f4493f0e3764e225b3f5a9a04700abb97258440e5"} Dec 09 15:44:49 crc kubenswrapper[4894]: I1209 15:44:49.481891 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jdblb" event={"ID":"2b7c2f69-58e9-4aad-9900-22befc1bc6c6","Type":"ContainerStarted","Data":"0c9d39fe8d1ac8087527542ae021a5926085aa4b059c12eee226de48322aeba3"} Dec 09 15:44:49 crc kubenswrapper[4894]: I1209 15:44:49.501539 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-f8648f98b-tjjzx" podStartSLOduration=2.501520169 podStartE2EDuration="2.501520169s" podCreationTimestamp="2025-12-09 15:44:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:44:49.500561694 +0000 UTC m=+783.819772413" watchObservedRunningTime="2025-12-09 15:44:49.501520169 +0000 UTC m=+783.820730838" Dec 09 15:44:49 crc kubenswrapper[4894]: I1209 15:44:49.636947 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d0f8c048-3892-4c12-a074-b7c4c5e85b87-memberlist\") pod \"speaker-8zntm\" (UID: \"d0f8c048-3892-4c12-a074-b7c4c5e85b87\") " pod="metallb-system/speaker-8zntm" Dec 09 15:44:49 crc kubenswrapper[4894]: I1209 15:44:49.652521 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/d0f8c048-3892-4c12-a074-b7c4c5e85b87-memberlist\") pod \"speaker-8zntm\" (UID: \"d0f8c048-3892-4c12-a074-b7c4c5e85b87\") " pod="metallb-system/speaker-8zntm" Dec 09 15:44:49 crc kubenswrapper[4894]: I1209 15:44:49.711209 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-8zntm" Dec 09 15:44:50 crc kubenswrapper[4894]: I1209 15:44:50.491268 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-8zntm" event={"ID":"d0f8c048-3892-4c12-a074-b7c4c5e85b87","Type":"ContainerStarted","Data":"6101d2bdbc3196df1f00fc4f3f2529b7c4e1a46d0cd9e644edf7bbde017bac1e"} Dec 09 15:44:50 crc kubenswrapper[4894]: I1209 15:44:50.491701 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-8zntm" event={"ID":"d0f8c048-3892-4c12-a074-b7c4c5e85b87","Type":"ContainerStarted","Data":"250ed6bea69ff84e986ffbb17b1b665bdf11e16385d340c543b4206f4d7f54ae"} Dec 09 15:44:50 crc kubenswrapper[4894]: I1209 15:44:50.491714 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-8zntm" event={"ID":"d0f8c048-3892-4c12-a074-b7c4c5e85b87","Type":"ContainerStarted","Data":"13799011e2d88f66d0fd95151bbb0034c4b013edcfea5ab6db6b025051ff3b92"} Dec 09 15:44:50 crc kubenswrapper[4894]: I1209 15:44:50.492353 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-8zntm" Dec 09 15:44:50 crc kubenswrapper[4894]: I1209 15:44:50.509524 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-8zntm" podStartSLOduration=3.509503071 podStartE2EDuration="3.509503071s" podCreationTimestamp="2025-12-09 15:44:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:44:50.506580184 +0000 UTC m=+784.825790853" watchObservedRunningTime="2025-12-09 15:44:50.509503071 +0000 UTC m=+784.828713740" Dec 09 15:44:55 crc kubenswrapper[4894]: I1209 15:44:55.532981 4894 generic.go:334] "Generic (PLEG): container finished" podID="2b7c2f69-58e9-4aad-9900-22befc1bc6c6" containerID="930572ba2e150f0a3888119a823be33a75575437ea7a21315fdea6883179f457" exitCode=0 Dec 09 15:44:55 crc kubenswrapper[4894]: I1209 15:44:55.534201 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jdblb" event={"ID":"2b7c2f69-58e9-4aad-9900-22befc1bc6c6","Type":"ContainerDied","Data":"930572ba2e150f0a3888119a823be33a75575437ea7a21315fdea6883179f457"} Dec 09 15:44:55 crc kubenswrapper[4894]: I1209 15:44:55.537174 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5dr97" event={"ID":"ba40556a-56db-491c-b0b0-e7b7f8ec1b6a","Type":"ContainerStarted","Data":"d4e9662e9d69e84bbd0a60d0c984e50223c0f0217b67c7c18d6d71ce8618f81d"} Dec 09 15:44:55 crc kubenswrapper[4894]: I1209 15:44:55.537312 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5dr97" Dec 09 15:44:55 crc kubenswrapper[4894]: I1209 15:44:55.585516 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5dr97" podStartSLOduration=2.197906137 podStartE2EDuration="8.585497147s" podCreationTimestamp="2025-12-09 15:44:47 +0000 UTC" firstStartedPulling="2025-12-09 15:44:48.956195311 +0000 UTC m=+783.275405980" lastFinishedPulling="2025-12-09 15:44:55.343786321 +0000 UTC m=+789.662996990" observedRunningTime="2025-12-09 15:44:55.584930652 +0000 UTC m=+789.904141321" watchObservedRunningTime="2025-12-09 15:44:55.585497147 +0000 UTC m=+789.904707816" Dec 09 15:44:56 crc kubenswrapper[4894]: I1209 15:44:56.545745 4894 generic.go:334] "Generic (PLEG): container finished" podID="2b7c2f69-58e9-4aad-9900-22befc1bc6c6" containerID="5379e2f691aa267b7e0a25bb4a5c20babd9c5c47dce19af93f7bf39ae63902b7" exitCode=0 Dec 09 15:44:56 crc kubenswrapper[4894]: I1209 15:44:56.545834 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jdblb" event={"ID":"2b7c2f69-58e9-4aad-9900-22befc1bc6c6","Type":"ContainerDied","Data":"5379e2f691aa267b7e0a25bb4a5c20babd9c5c47dce19af93f7bf39ae63902b7"} Dec 09 15:44:57 crc kubenswrapper[4894]: I1209 15:44:57.554306 4894 generic.go:334] "Generic (PLEG): container finished" podID="2b7c2f69-58e9-4aad-9900-22befc1bc6c6" containerID="4a1cd5a6d29869ddf621d3a2005d08526438174c71baf6e3349f12d8d5f6bdb9" exitCode=0 Dec 09 15:44:57 crc kubenswrapper[4894]: I1209 15:44:57.554374 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jdblb" event={"ID":"2b7c2f69-58e9-4aad-9900-22befc1bc6c6","Type":"ContainerDied","Data":"4a1cd5a6d29869ddf621d3a2005d08526438174c71baf6e3349f12d8d5f6bdb9"} Dec 09 15:44:58 crc kubenswrapper[4894]: I1209 15:44:58.248945 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-f8648f98b-tjjzx" Dec 09 15:44:58 crc kubenswrapper[4894]: I1209 15:44:58.567269 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jdblb" event={"ID":"2b7c2f69-58e9-4aad-9900-22befc1bc6c6","Type":"ContainerStarted","Data":"ed09bf8d8f803ab313aa7781c2ff170f88f1659726ba1148815df30863e2c232"} Dec 09 15:44:58 crc kubenswrapper[4894]: I1209 15:44:58.567319 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jdblb" event={"ID":"2b7c2f69-58e9-4aad-9900-22befc1bc6c6","Type":"ContainerStarted","Data":"b38a548a12ab6d0fe5e11057d4feda0a16c308eae23007deb3c1c8d8112c5162"} Dec 09 15:44:58 crc kubenswrapper[4894]: I1209 15:44:58.567334 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jdblb" event={"ID":"2b7c2f69-58e9-4aad-9900-22befc1bc6c6","Type":"ContainerStarted","Data":"fc1d7f67b45ed86d1821f1702368fba51535818dfc09e1ab5735c6ceffb17c5a"} Dec 09 15:44:58 crc kubenswrapper[4894]: I1209 15:44:58.567346 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jdblb" event={"ID":"2b7c2f69-58e9-4aad-9900-22befc1bc6c6","Type":"ContainerStarted","Data":"f1975ace0068cb103d194207456087b057431e2f4e50b2510da4aaafa8b7c836"} Dec 09 15:44:58 crc kubenswrapper[4894]: I1209 15:44:58.567358 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jdblb" event={"ID":"2b7c2f69-58e9-4aad-9900-22befc1bc6c6","Type":"ContainerStarted","Data":"7314f3d87848adf634be5474059666a672287f5dd3e4209a1ea7eebb8673406a"} Dec 09 15:44:59 crc kubenswrapper[4894]: I1209 15:44:59.583877 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-jdblb" event={"ID":"2b7c2f69-58e9-4aad-9900-22befc1bc6c6","Type":"ContainerStarted","Data":"8870853fec5372db792f663aaa57d2f0c806f888f8bcb22fd9615b8394fabd84"} Dec 09 15:44:59 crc kubenswrapper[4894]: I1209 15:44:59.584167 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-jdblb" Dec 09 15:44:59 crc kubenswrapper[4894]: I1209 15:44:59.606612 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-jdblb" podStartSLOduration=6.1292916 podStartE2EDuration="12.60659138s" podCreationTimestamp="2025-12-09 15:44:47 +0000 UTC" firstStartedPulling="2025-12-09 15:44:48.835880812 +0000 UTC m=+783.155091471" lastFinishedPulling="2025-12-09 15:44:55.313180592 +0000 UTC m=+789.632391251" observedRunningTime="2025-12-09 15:44:59.605969503 +0000 UTC m=+793.925180182" watchObservedRunningTime="2025-12-09 15:44:59.60659138 +0000 UTC m=+793.925802059" Dec 09 15:44:59 crc kubenswrapper[4894]: I1209 15:44:59.714400 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-8zntm" Dec 09 15:45:00 crc kubenswrapper[4894]: I1209 15:45:00.164402 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421585-wrfw9"] Dec 09 15:45:00 crc kubenswrapper[4894]: I1209 15:45:00.165157 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421585-wrfw9" Dec 09 15:45:00 crc kubenswrapper[4894]: I1209 15:45:00.167067 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 15:45:00 crc kubenswrapper[4894]: I1209 15:45:00.167257 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 15:45:00 crc kubenswrapper[4894]: I1209 15:45:00.179175 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421585-wrfw9"] Dec 09 15:45:00 crc kubenswrapper[4894]: I1209 15:45:00.302348 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snc8c\" (UniqueName: \"kubernetes.io/projected/67604b04-833c-4a08-a809-860c3fdbf802-kube-api-access-snc8c\") pod \"collect-profiles-29421585-wrfw9\" (UID: \"67604b04-833c-4a08-a809-860c3fdbf802\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421585-wrfw9" Dec 09 15:45:00 crc kubenswrapper[4894]: I1209 15:45:00.302716 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/67604b04-833c-4a08-a809-860c3fdbf802-config-volume\") pod \"collect-profiles-29421585-wrfw9\" (UID: \"67604b04-833c-4a08-a809-860c3fdbf802\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421585-wrfw9" Dec 09 15:45:00 crc kubenswrapper[4894]: I1209 15:45:00.302801 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/67604b04-833c-4a08-a809-860c3fdbf802-secret-volume\") pod \"collect-profiles-29421585-wrfw9\" (UID: \"67604b04-833c-4a08-a809-860c3fdbf802\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421585-wrfw9" Dec 09 15:45:00 crc kubenswrapper[4894]: I1209 15:45:00.403627 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/67604b04-833c-4a08-a809-860c3fdbf802-secret-volume\") pod \"collect-profiles-29421585-wrfw9\" (UID: \"67604b04-833c-4a08-a809-860c3fdbf802\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421585-wrfw9" Dec 09 15:45:00 crc kubenswrapper[4894]: I1209 15:45:00.403731 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snc8c\" (UniqueName: \"kubernetes.io/projected/67604b04-833c-4a08-a809-860c3fdbf802-kube-api-access-snc8c\") pod \"collect-profiles-29421585-wrfw9\" (UID: \"67604b04-833c-4a08-a809-860c3fdbf802\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421585-wrfw9" Dec 09 15:45:00 crc kubenswrapper[4894]: I1209 15:45:00.403773 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/67604b04-833c-4a08-a809-860c3fdbf802-config-volume\") pod \"collect-profiles-29421585-wrfw9\" (UID: \"67604b04-833c-4a08-a809-860c3fdbf802\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421585-wrfw9" Dec 09 15:45:00 crc kubenswrapper[4894]: I1209 15:45:00.405274 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/67604b04-833c-4a08-a809-860c3fdbf802-config-volume\") pod \"collect-profiles-29421585-wrfw9\" (UID: \"67604b04-833c-4a08-a809-860c3fdbf802\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421585-wrfw9" Dec 09 15:45:00 crc kubenswrapper[4894]: I1209 15:45:00.410414 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/67604b04-833c-4a08-a809-860c3fdbf802-secret-volume\") pod \"collect-profiles-29421585-wrfw9\" (UID: \"67604b04-833c-4a08-a809-860c3fdbf802\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421585-wrfw9" Dec 09 15:45:00 crc kubenswrapper[4894]: I1209 15:45:00.420971 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snc8c\" (UniqueName: \"kubernetes.io/projected/67604b04-833c-4a08-a809-860c3fdbf802-kube-api-access-snc8c\") pod \"collect-profiles-29421585-wrfw9\" (UID: \"67604b04-833c-4a08-a809-860c3fdbf802\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421585-wrfw9" Dec 09 15:45:00 crc kubenswrapper[4894]: I1209 15:45:00.487826 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421585-wrfw9" Dec 09 15:45:00 crc kubenswrapper[4894]: I1209 15:45:00.934114 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421585-wrfw9"] Dec 09 15:45:01 crc kubenswrapper[4894]: I1209 15:45:01.599877 4894 generic.go:334] "Generic (PLEG): container finished" podID="67604b04-833c-4a08-a809-860c3fdbf802" containerID="03d5d47017f84d6610e92b5ac3f39af962e1fb0802f4fe13bd6b393b4d0aeff7" exitCode=0 Dec 09 15:45:01 crc kubenswrapper[4894]: I1209 15:45:01.600004 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421585-wrfw9" event={"ID":"67604b04-833c-4a08-a809-860c3fdbf802","Type":"ContainerDied","Data":"03d5d47017f84d6610e92b5ac3f39af962e1fb0802f4fe13bd6b393b4d0aeff7"} Dec 09 15:45:01 crc kubenswrapper[4894]: I1209 15:45:01.600327 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421585-wrfw9" event={"ID":"67604b04-833c-4a08-a809-860c3fdbf802","Type":"ContainerStarted","Data":"40557f894ee97db50cca18f336b6d29900aa738f8bc3e6771af09ff84029192b"} Dec 09 15:45:02 crc kubenswrapper[4894]: I1209 15:45:02.508512 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-hjbcx"] Dec 09 15:45:02 crc kubenswrapper[4894]: I1209 15:45:02.511105 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hjbcx" Dec 09 15:45:02 crc kubenswrapper[4894]: I1209 15:45:02.518789 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Dec 09 15:45:02 crc kubenswrapper[4894]: I1209 15:45:02.518833 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-pjq82" Dec 09 15:45:02 crc kubenswrapper[4894]: I1209 15:45:02.526440 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Dec 09 15:45:02 crc kubenswrapper[4894]: I1209 15:45:02.529461 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hjbcx"] Dec 09 15:45:02 crc kubenswrapper[4894]: I1209 15:45:02.532926 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bgdt8\" (UniqueName: \"kubernetes.io/projected/e1694082-6d08-47ff-8bdb-2d5c9f48b9fa-kube-api-access-bgdt8\") pod \"openstack-operator-index-hjbcx\" (UID: \"e1694082-6d08-47ff-8bdb-2d5c9f48b9fa\") " pod="openstack-operators/openstack-operator-index-hjbcx" Dec 09 15:45:02 crc kubenswrapper[4894]: I1209 15:45:02.633817 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bgdt8\" (UniqueName: \"kubernetes.io/projected/e1694082-6d08-47ff-8bdb-2d5c9f48b9fa-kube-api-access-bgdt8\") pod \"openstack-operator-index-hjbcx\" (UID: \"e1694082-6d08-47ff-8bdb-2d5c9f48b9fa\") " pod="openstack-operators/openstack-operator-index-hjbcx" Dec 09 15:45:02 crc kubenswrapper[4894]: I1209 15:45:02.655528 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bgdt8\" (UniqueName: \"kubernetes.io/projected/e1694082-6d08-47ff-8bdb-2d5c9f48b9fa-kube-api-access-bgdt8\") pod \"openstack-operator-index-hjbcx\" (UID: \"e1694082-6d08-47ff-8bdb-2d5c9f48b9fa\") " pod="openstack-operators/openstack-operator-index-hjbcx" Dec 09 15:45:02 crc kubenswrapper[4894]: I1209 15:45:02.830908 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hjbcx" Dec 09 15:45:03 crc kubenswrapper[4894]: I1209 15:45:02.835730 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421585-wrfw9" Dec 09 15:45:03 crc kubenswrapper[4894]: I1209 15:45:02.836941 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/67604b04-833c-4a08-a809-860c3fdbf802-config-volume\") pod \"67604b04-833c-4a08-a809-860c3fdbf802\" (UID: \"67604b04-833c-4a08-a809-860c3fdbf802\") " Dec 09 15:45:03 crc kubenswrapper[4894]: I1209 15:45:02.836977 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/67604b04-833c-4a08-a809-860c3fdbf802-secret-volume\") pod \"67604b04-833c-4a08-a809-860c3fdbf802\" (UID: \"67604b04-833c-4a08-a809-860c3fdbf802\") " Dec 09 15:45:03 crc kubenswrapper[4894]: I1209 15:45:02.837083 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-snc8c\" (UniqueName: \"kubernetes.io/projected/67604b04-833c-4a08-a809-860c3fdbf802-kube-api-access-snc8c\") pod \"67604b04-833c-4a08-a809-860c3fdbf802\" (UID: \"67604b04-833c-4a08-a809-860c3fdbf802\") " Dec 09 15:45:03 crc kubenswrapper[4894]: I1209 15:45:02.837475 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/67604b04-833c-4a08-a809-860c3fdbf802-config-volume" (OuterVolumeSpecName: "config-volume") pod "67604b04-833c-4a08-a809-860c3fdbf802" (UID: "67604b04-833c-4a08-a809-860c3fdbf802"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:45:03 crc kubenswrapper[4894]: I1209 15:45:02.937944 4894 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/67604b04-833c-4a08-a809-860c3fdbf802-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 15:45:03 crc kubenswrapper[4894]: I1209 15:45:03.149873 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/67604b04-833c-4a08-a809-860c3fdbf802-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "67604b04-833c-4a08-a809-860c3fdbf802" (UID: "67604b04-833c-4a08-a809-860c3fdbf802"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:45:03 crc kubenswrapper[4894]: I1209 15:45:03.152991 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67604b04-833c-4a08-a809-860c3fdbf802-kube-api-access-snc8c" (OuterVolumeSpecName: "kube-api-access-snc8c") pod "67604b04-833c-4a08-a809-860c3fdbf802" (UID: "67604b04-833c-4a08-a809-860c3fdbf802"). InnerVolumeSpecName "kube-api-access-snc8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:45:03 crc kubenswrapper[4894]: I1209 15:45:03.252506 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-snc8c\" (UniqueName: \"kubernetes.io/projected/67604b04-833c-4a08-a809-860c3fdbf802-kube-api-access-snc8c\") on node \"crc\" DevicePath \"\"" Dec 09 15:45:03 crc kubenswrapper[4894]: I1209 15:45:03.252548 4894 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/67604b04-833c-4a08-a809-860c3fdbf802-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 15:45:03 crc kubenswrapper[4894]: I1209 15:45:03.365119 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-hjbcx"] Dec 09 15:45:03 crc kubenswrapper[4894]: W1209 15:45:03.367711 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode1694082_6d08_47ff_8bdb_2d5c9f48b9fa.slice/crio-bac4527f482fd7c50614b1a6bff1126e25e42f6c2f856f5a3a5740118d3808c5 WatchSource:0}: Error finding container bac4527f482fd7c50614b1a6bff1126e25e42f6c2f856f5a3a5740118d3808c5: Status 404 returned error can't find the container with id bac4527f482fd7c50614b1a6bff1126e25e42f6c2f856f5a3a5740118d3808c5 Dec 09 15:45:03 crc kubenswrapper[4894]: I1209 15:45:03.613312 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hjbcx" event={"ID":"e1694082-6d08-47ff-8bdb-2d5c9f48b9fa","Type":"ContainerStarted","Data":"bac4527f482fd7c50614b1a6bff1126e25e42f6c2f856f5a3a5740118d3808c5"} Dec 09 15:45:03 crc kubenswrapper[4894]: I1209 15:45:03.614992 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421585-wrfw9" event={"ID":"67604b04-833c-4a08-a809-860c3fdbf802","Type":"ContainerDied","Data":"40557f894ee97db50cca18f336b6d29900aa738f8bc3e6771af09ff84029192b"} Dec 09 15:45:03 crc kubenswrapper[4894]: I1209 15:45:03.615025 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="40557f894ee97db50cca18f336b6d29900aa738f8bc3e6771af09ff84029192b" Dec 09 15:45:03 crc kubenswrapper[4894]: I1209 15:45:03.615045 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421585-wrfw9" Dec 09 15:45:03 crc kubenswrapper[4894]: I1209 15:45:03.736211 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-jdblb" Dec 09 15:45:03 crc kubenswrapper[4894]: I1209 15:45:03.777119 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-jdblb" Dec 09 15:45:04 crc kubenswrapper[4894]: I1209 15:45:04.622604 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hjbcx" event={"ID":"e1694082-6d08-47ff-8bdb-2d5c9f48b9fa","Type":"ContainerStarted","Data":"98afe826dd38391d03e9c936dfd810eb4363158c43bef30f7f4ed1f4e5b318eb"} Dec 09 15:45:04 crc kubenswrapper[4894]: I1209 15:45:04.667373 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-hjbcx" podStartSLOduration=1.692802544 podStartE2EDuration="2.667348542s" podCreationTimestamp="2025-12-09 15:45:02 +0000 UTC" firstStartedPulling="2025-12-09 15:45:03.36984431 +0000 UTC m=+797.689054979" lastFinishedPulling="2025-12-09 15:45:04.344390318 +0000 UTC m=+798.663600977" observedRunningTime="2025-12-09 15:45:04.664584099 +0000 UTC m=+798.983794768" watchObservedRunningTime="2025-12-09 15:45:04.667348542 +0000 UTC m=+798.986559251" Dec 09 15:45:05 crc kubenswrapper[4894]: I1209 15:45:05.702443 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-hjbcx"] Dec 09 15:45:06 crc kubenswrapper[4894]: I1209 15:45:06.300428 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-m8kvv"] Dec 09 15:45:06 crc kubenswrapper[4894]: E1209 15:45:06.301522 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67604b04-833c-4a08-a809-860c3fdbf802" containerName="collect-profiles" Dec 09 15:45:06 crc kubenswrapper[4894]: I1209 15:45:06.301556 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="67604b04-833c-4a08-a809-860c3fdbf802" containerName="collect-profiles" Dec 09 15:45:06 crc kubenswrapper[4894]: I1209 15:45:06.301757 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="67604b04-833c-4a08-a809-860c3fdbf802" containerName="collect-profiles" Dec 09 15:45:06 crc kubenswrapper[4894]: I1209 15:45:06.302367 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-m8kvv" Dec 09 15:45:06 crc kubenswrapper[4894]: I1209 15:45:06.309037 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-m8kvv"] Dec 09 15:45:06 crc kubenswrapper[4894]: I1209 15:45:06.391963 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p9646\" (UniqueName: \"kubernetes.io/projected/cfb263e2-5739-43ed-965a-575f4cb83652-kube-api-access-p9646\") pod \"openstack-operator-index-m8kvv\" (UID: \"cfb263e2-5739-43ed-965a-575f4cb83652\") " pod="openstack-operators/openstack-operator-index-m8kvv" Dec 09 15:45:06 crc kubenswrapper[4894]: I1209 15:45:06.493557 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p9646\" (UniqueName: \"kubernetes.io/projected/cfb263e2-5739-43ed-965a-575f4cb83652-kube-api-access-p9646\") pod \"openstack-operator-index-m8kvv\" (UID: \"cfb263e2-5739-43ed-965a-575f4cb83652\") " pod="openstack-operators/openstack-operator-index-m8kvv" Dec 09 15:45:06 crc kubenswrapper[4894]: I1209 15:45:06.524939 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p9646\" (UniqueName: \"kubernetes.io/projected/cfb263e2-5739-43ed-965a-575f4cb83652-kube-api-access-p9646\") pod \"openstack-operator-index-m8kvv\" (UID: \"cfb263e2-5739-43ed-965a-575f4cb83652\") " pod="openstack-operators/openstack-operator-index-m8kvv" Dec 09 15:45:06 crc kubenswrapper[4894]: I1209 15:45:06.617861 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-m8kvv" Dec 09 15:45:06 crc kubenswrapper[4894]: I1209 15:45:06.632340 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-hjbcx" podUID="e1694082-6d08-47ff-8bdb-2d5c9f48b9fa" containerName="registry-server" containerID="cri-o://98afe826dd38391d03e9c936dfd810eb4363158c43bef30f7f4ed1f4e5b318eb" gracePeriod=2 Dec 09 15:45:06 crc kubenswrapper[4894]: I1209 15:45:06.998208 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hjbcx" Dec 09 15:45:07 crc kubenswrapper[4894]: I1209 15:45:07.063786 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-m8kvv"] Dec 09 15:45:07 crc kubenswrapper[4894]: W1209 15:45:07.076011 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfb263e2_5739_43ed_965a_575f4cb83652.slice/crio-0fd67923eabfb0ffdd181267cdcf5525577501ecc434c33a131e06a17c25708d WatchSource:0}: Error finding container 0fd67923eabfb0ffdd181267cdcf5525577501ecc434c33a131e06a17c25708d: Status 404 returned error can't find the container with id 0fd67923eabfb0ffdd181267cdcf5525577501ecc434c33a131e06a17c25708d Dec 09 15:45:07 crc kubenswrapper[4894]: I1209 15:45:07.101459 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bgdt8\" (UniqueName: \"kubernetes.io/projected/e1694082-6d08-47ff-8bdb-2d5c9f48b9fa-kube-api-access-bgdt8\") pod \"e1694082-6d08-47ff-8bdb-2d5c9f48b9fa\" (UID: \"e1694082-6d08-47ff-8bdb-2d5c9f48b9fa\") " Dec 09 15:45:07 crc kubenswrapper[4894]: I1209 15:45:07.108524 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e1694082-6d08-47ff-8bdb-2d5c9f48b9fa-kube-api-access-bgdt8" (OuterVolumeSpecName: "kube-api-access-bgdt8") pod "e1694082-6d08-47ff-8bdb-2d5c9f48b9fa" (UID: "e1694082-6d08-47ff-8bdb-2d5c9f48b9fa"). InnerVolumeSpecName "kube-api-access-bgdt8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:45:07 crc kubenswrapper[4894]: I1209 15:45:07.204123 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bgdt8\" (UniqueName: \"kubernetes.io/projected/e1694082-6d08-47ff-8bdb-2d5c9f48b9fa-kube-api-access-bgdt8\") on node \"crc\" DevicePath \"\"" Dec 09 15:45:07 crc kubenswrapper[4894]: I1209 15:45:07.653778 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-m8kvv" event={"ID":"cfb263e2-5739-43ed-965a-575f4cb83652","Type":"ContainerStarted","Data":"0fd67923eabfb0ffdd181267cdcf5525577501ecc434c33a131e06a17c25708d"} Dec 09 15:45:07 crc kubenswrapper[4894]: I1209 15:45:07.655580 4894 generic.go:334] "Generic (PLEG): container finished" podID="e1694082-6d08-47ff-8bdb-2d5c9f48b9fa" containerID="98afe826dd38391d03e9c936dfd810eb4363158c43bef30f7f4ed1f4e5b318eb" exitCode=0 Dec 09 15:45:07 crc kubenswrapper[4894]: I1209 15:45:07.655622 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hjbcx" event={"ID":"e1694082-6d08-47ff-8bdb-2d5c9f48b9fa","Type":"ContainerDied","Data":"98afe826dd38391d03e9c936dfd810eb4363158c43bef30f7f4ed1f4e5b318eb"} Dec 09 15:45:07 crc kubenswrapper[4894]: I1209 15:45:07.655666 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-hjbcx" event={"ID":"e1694082-6d08-47ff-8bdb-2d5c9f48b9fa","Type":"ContainerDied","Data":"bac4527f482fd7c50614b1a6bff1126e25e42f6c2f856f5a3a5740118d3808c5"} Dec 09 15:45:07 crc kubenswrapper[4894]: I1209 15:45:07.655686 4894 scope.go:117] "RemoveContainer" containerID="98afe826dd38391d03e9c936dfd810eb4363158c43bef30f7f4ed1f4e5b318eb" Dec 09 15:45:07 crc kubenswrapper[4894]: I1209 15:45:07.655693 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-hjbcx" Dec 09 15:45:07 crc kubenswrapper[4894]: I1209 15:45:07.673522 4894 scope.go:117] "RemoveContainer" containerID="98afe826dd38391d03e9c936dfd810eb4363158c43bef30f7f4ed1f4e5b318eb" Dec 09 15:45:07 crc kubenswrapper[4894]: E1209 15:45:07.674307 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"98afe826dd38391d03e9c936dfd810eb4363158c43bef30f7f4ed1f4e5b318eb\": container with ID starting with 98afe826dd38391d03e9c936dfd810eb4363158c43bef30f7f4ed1f4e5b318eb not found: ID does not exist" containerID="98afe826dd38391d03e9c936dfd810eb4363158c43bef30f7f4ed1f4e5b318eb" Dec 09 15:45:07 crc kubenswrapper[4894]: I1209 15:45:07.674763 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"98afe826dd38391d03e9c936dfd810eb4363158c43bef30f7f4ed1f4e5b318eb"} err="failed to get container status \"98afe826dd38391d03e9c936dfd810eb4363158c43bef30f7f4ed1f4e5b318eb\": rpc error: code = NotFound desc = could not find container \"98afe826dd38391d03e9c936dfd810eb4363158c43bef30f7f4ed1f4e5b318eb\": container with ID starting with 98afe826dd38391d03e9c936dfd810eb4363158c43bef30f7f4ed1f4e5b318eb not found: ID does not exist" Dec 09 15:45:07 crc kubenswrapper[4894]: I1209 15:45:07.695196 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-hjbcx"] Dec 09 15:45:07 crc kubenswrapper[4894]: I1209 15:45:07.698997 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-hjbcx"] Dec 09 15:45:08 crc kubenswrapper[4894]: I1209 15:45:08.114174 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e1694082-6d08-47ff-8bdb-2d5c9f48b9fa" path="/var/lib/kubelet/pods/e1694082-6d08-47ff-8bdb-2d5c9f48b9fa/volumes" Dec 09 15:45:08 crc kubenswrapper[4894]: I1209 15:45:08.736312 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-jdblb" Dec 09 15:45:08 crc kubenswrapper[4894]: I1209 15:45:08.747000 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-7fcb986d4-5dr97" Dec 09 15:45:09 crc kubenswrapper[4894]: I1209 15:45:09.670891 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-m8kvv" event={"ID":"cfb263e2-5739-43ed-965a-575f4cb83652","Type":"ContainerStarted","Data":"10bdd4112f896e24f037fe3fc480a34c67ac03775e6e6e925063e41061319cf7"} Dec 09 15:45:09 crc kubenswrapper[4894]: I1209 15:45:09.692399 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-m8kvv" podStartSLOduration=2.037382443 podStartE2EDuration="3.69236834s" podCreationTimestamp="2025-12-09 15:45:06 +0000 UTC" firstStartedPulling="2025-12-09 15:45:07.081819156 +0000 UTC m=+801.401029825" lastFinishedPulling="2025-12-09 15:45:08.736805053 +0000 UTC m=+803.056015722" observedRunningTime="2025-12-09 15:45:09.687399428 +0000 UTC m=+804.006610097" watchObservedRunningTime="2025-12-09 15:45:09.69236834 +0000 UTC m=+804.011579019" Dec 09 15:45:16 crc kubenswrapper[4894]: I1209 15:45:16.618117 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-m8kvv" Dec 09 15:45:16 crc kubenswrapper[4894]: I1209 15:45:16.619818 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-m8kvv" Dec 09 15:45:16 crc kubenswrapper[4894]: I1209 15:45:16.662653 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-m8kvv" Dec 09 15:45:16 crc kubenswrapper[4894]: I1209 15:45:16.742051 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-m8kvv" Dec 09 15:45:18 crc kubenswrapper[4894]: I1209 15:45:18.939093 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b"] Dec 09 15:45:18 crc kubenswrapper[4894]: E1209 15:45:18.939467 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e1694082-6d08-47ff-8bdb-2d5c9f48b9fa" containerName="registry-server" Dec 09 15:45:18 crc kubenswrapper[4894]: I1209 15:45:18.939488 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e1694082-6d08-47ff-8bdb-2d5c9f48b9fa" containerName="registry-server" Dec 09 15:45:18 crc kubenswrapper[4894]: I1209 15:45:18.939729 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e1694082-6d08-47ff-8bdb-2d5c9f48b9fa" containerName="registry-server" Dec 09 15:45:18 crc kubenswrapper[4894]: I1209 15:45:18.941388 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b" Dec 09 15:45:18 crc kubenswrapper[4894]: I1209 15:45:18.943230 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-8gq9s" Dec 09 15:45:18 crc kubenswrapper[4894]: I1209 15:45:18.947531 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b"] Dec 09 15:45:19 crc kubenswrapper[4894]: I1209 15:45:19.067985 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvn6t\" (UniqueName: \"kubernetes.io/projected/caf14a3e-1621-42fc-9f4f-95d64b325062-kube-api-access-xvn6t\") pod \"a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b\" (UID: \"caf14a3e-1621-42fc-9f4f-95d64b325062\") " pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b" Dec 09 15:45:19 crc kubenswrapper[4894]: I1209 15:45:19.068335 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/caf14a3e-1621-42fc-9f4f-95d64b325062-util\") pod \"a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b\" (UID: \"caf14a3e-1621-42fc-9f4f-95d64b325062\") " pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b" Dec 09 15:45:19 crc kubenswrapper[4894]: I1209 15:45:19.068372 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/caf14a3e-1621-42fc-9f4f-95d64b325062-bundle\") pod \"a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b\" (UID: \"caf14a3e-1621-42fc-9f4f-95d64b325062\") " pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b" Dec 09 15:45:19 crc kubenswrapper[4894]: I1209 15:45:19.170145 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvn6t\" (UniqueName: \"kubernetes.io/projected/caf14a3e-1621-42fc-9f4f-95d64b325062-kube-api-access-xvn6t\") pod \"a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b\" (UID: \"caf14a3e-1621-42fc-9f4f-95d64b325062\") " pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b" Dec 09 15:45:19 crc kubenswrapper[4894]: I1209 15:45:19.170257 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/caf14a3e-1621-42fc-9f4f-95d64b325062-util\") pod \"a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b\" (UID: \"caf14a3e-1621-42fc-9f4f-95d64b325062\") " pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b" Dec 09 15:45:19 crc kubenswrapper[4894]: I1209 15:45:19.170328 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/caf14a3e-1621-42fc-9f4f-95d64b325062-bundle\") pod \"a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b\" (UID: \"caf14a3e-1621-42fc-9f4f-95d64b325062\") " pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b" Dec 09 15:45:19 crc kubenswrapper[4894]: I1209 15:45:19.170773 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/caf14a3e-1621-42fc-9f4f-95d64b325062-util\") pod \"a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b\" (UID: \"caf14a3e-1621-42fc-9f4f-95d64b325062\") " pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b" Dec 09 15:45:19 crc kubenswrapper[4894]: I1209 15:45:19.170929 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/caf14a3e-1621-42fc-9f4f-95d64b325062-bundle\") pod \"a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b\" (UID: \"caf14a3e-1621-42fc-9f4f-95d64b325062\") " pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b" Dec 09 15:45:19 crc kubenswrapper[4894]: I1209 15:45:19.193434 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvn6t\" (UniqueName: \"kubernetes.io/projected/caf14a3e-1621-42fc-9f4f-95d64b325062-kube-api-access-xvn6t\") pod \"a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b\" (UID: \"caf14a3e-1621-42fc-9f4f-95d64b325062\") " pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b" Dec 09 15:45:19 crc kubenswrapper[4894]: I1209 15:45:19.259883 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b" Dec 09 15:45:19 crc kubenswrapper[4894]: I1209 15:45:19.664168 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b"] Dec 09 15:45:19 crc kubenswrapper[4894]: I1209 15:45:19.734365 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b" event={"ID":"caf14a3e-1621-42fc-9f4f-95d64b325062","Type":"ContainerStarted","Data":"0e61fddcf9e2b780ef33492b50d62f607b186bf18a3e6c01b06179e65dae69dd"} Dec 09 15:45:20 crc kubenswrapper[4894]: I1209 15:45:20.742526 4894 generic.go:334] "Generic (PLEG): container finished" podID="caf14a3e-1621-42fc-9f4f-95d64b325062" containerID="eb3f08803c04ccb05a75cba1a154f622e24eb24e65abbeaa60224adea23de979" exitCode=0 Dec 09 15:45:20 crc kubenswrapper[4894]: I1209 15:45:20.742629 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b" event={"ID":"caf14a3e-1621-42fc-9f4f-95d64b325062","Type":"ContainerDied","Data":"eb3f08803c04ccb05a75cba1a154f622e24eb24e65abbeaa60224adea23de979"} Dec 09 15:45:21 crc kubenswrapper[4894]: I1209 15:45:21.751358 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b" event={"ID":"caf14a3e-1621-42fc-9f4f-95d64b325062","Type":"ContainerStarted","Data":"c959e11c07727291f01e90ec79118013294287ac3e57c7f8a5d73aa9f9170022"} Dec 09 15:45:22 crc kubenswrapper[4894]: I1209 15:45:22.761032 4894 generic.go:334] "Generic (PLEG): container finished" podID="caf14a3e-1621-42fc-9f4f-95d64b325062" containerID="c959e11c07727291f01e90ec79118013294287ac3e57c7f8a5d73aa9f9170022" exitCode=0 Dec 09 15:45:22 crc kubenswrapper[4894]: I1209 15:45:22.761377 4894 generic.go:334] "Generic (PLEG): container finished" podID="caf14a3e-1621-42fc-9f4f-95d64b325062" containerID="19391c5c36583ab52d145f0c64b74eda82959a101b2547c9278d5bfb180a2d77" exitCode=0 Dec 09 15:45:22 crc kubenswrapper[4894]: I1209 15:45:22.761396 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b" event={"ID":"caf14a3e-1621-42fc-9f4f-95d64b325062","Type":"ContainerDied","Data":"c959e11c07727291f01e90ec79118013294287ac3e57c7f8a5d73aa9f9170022"} Dec 09 15:45:22 crc kubenswrapper[4894]: I1209 15:45:22.761420 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b" event={"ID":"caf14a3e-1621-42fc-9f4f-95d64b325062","Type":"ContainerDied","Data":"19391c5c36583ab52d145f0c64b74eda82959a101b2547c9278d5bfb180a2d77"} Dec 09 15:45:24 crc kubenswrapper[4894]: I1209 15:45:24.002268 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b" Dec 09 15:45:24 crc kubenswrapper[4894]: I1209 15:45:24.133753 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/caf14a3e-1621-42fc-9f4f-95d64b325062-bundle\") pod \"caf14a3e-1621-42fc-9f4f-95d64b325062\" (UID: \"caf14a3e-1621-42fc-9f4f-95d64b325062\") " Dec 09 15:45:24 crc kubenswrapper[4894]: I1209 15:45:24.133818 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvn6t\" (UniqueName: \"kubernetes.io/projected/caf14a3e-1621-42fc-9f4f-95d64b325062-kube-api-access-xvn6t\") pod \"caf14a3e-1621-42fc-9f4f-95d64b325062\" (UID: \"caf14a3e-1621-42fc-9f4f-95d64b325062\") " Dec 09 15:45:24 crc kubenswrapper[4894]: I1209 15:45:24.133895 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/caf14a3e-1621-42fc-9f4f-95d64b325062-util\") pod \"caf14a3e-1621-42fc-9f4f-95d64b325062\" (UID: \"caf14a3e-1621-42fc-9f4f-95d64b325062\") " Dec 09 15:45:24 crc kubenswrapper[4894]: I1209 15:45:24.135001 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caf14a3e-1621-42fc-9f4f-95d64b325062-bundle" (OuterVolumeSpecName: "bundle") pod "caf14a3e-1621-42fc-9f4f-95d64b325062" (UID: "caf14a3e-1621-42fc-9f4f-95d64b325062"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:45:24 crc kubenswrapper[4894]: I1209 15:45:24.143706 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caf14a3e-1621-42fc-9f4f-95d64b325062-kube-api-access-xvn6t" (OuterVolumeSpecName: "kube-api-access-xvn6t") pod "caf14a3e-1621-42fc-9f4f-95d64b325062" (UID: "caf14a3e-1621-42fc-9f4f-95d64b325062"). InnerVolumeSpecName "kube-api-access-xvn6t". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:45:24 crc kubenswrapper[4894]: I1209 15:45:24.150746 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/caf14a3e-1621-42fc-9f4f-95d64b325062-util" (OuterVolumeSpecName: "util") pod "caf14a3e-1621-42fc-9f4f-95d64b325062" (UID: "caf14a3e-1621-42fc-9f4f-95d64b325062"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:45:24 crc kubenswrapper[4894]: I1209 15:45:24.235079 4894 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/caf14a3e-1621-42fc-9f4f-95d64b325062-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:45:24 crc kubenswrapper[4894]: I1209 15:45:24.235117 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvn6t\" (UniqueName: \"kubernetes.io/projected/caf14a3e-1621-42fc-9f4f-95d64b325062-kube-api-access-xvn6t\") on node \"crc\" DevicePath \"\"" Dec 09 15:45:24 crc kubenswrapper[4894]: I1209 15:45:24.235130 4894 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/caf14a3e-1621-42fc-9f4f-95d64b325062-util\") on node \"crc\" DevicePath \"\"" Dec 09 15:45:24 crc kubenswrapper[4894]: I1209 15:45:24.773185 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b" event={"ID":"caf14a3e-1621-42fc-9f4f-95d64b325062","Type":"ContainerDied","Data":"0e61fddcf9e2b780ef33492b50d62f607b186bf18a3e6c01b06179e65dae69dd"} Dec 09 15:45:24 crc kubenswrapper[4894]: I1209 15:45:24.773222 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e61fddcf9e2b780ef33492b50d62f607b186bf18a3e6c01b06179e65dae69dd" Dec 09 15:45:24 crc kubenswrapper[4894]: I1209 15:45:24.773561 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b" Dec 09 15:45:31 crc kubenswrapper[4894]: I1209 15:45:31.111941 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7979d445b4-d7b6w"] Dec 09 15:45:31 crc kubenswrapper[4894]: E1209 15:45:31.112792 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caf14a3e-1621-42fc-9f4f-95d64b325062" containerName="extract" Dec 09 15:45:31 crc kubenswrapper[4894]: I1209 15:45:31.112804 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="caf14a3e-1621-42fc-9f4f-95d64b325062" containerName="extract" Dec 09 15:45:31 crc kubenswrapper[4894]: E1209 15:45:31.112818 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caf14a3e-1621-42fc-9f4f-95d64b325062" containerName="pull" Dec 09 15:45:31 crc kubenswrapper[4894]: I1209 15:45:31.112824 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="caf14a3e-1621-42fc-9f4f-95d64b325062" containerName="pull" Dec 09 15:45:31 crc kubenswrapper[4894]: E1209 15:45:31.112834 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caf14a3e-1621-42fc-9f4f-95d64b325062" containerName="util" Dec 09 15:45:31 crc kubenswrapper[4894]: I1209 15:45:31.112841 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="caf14a3e-1621-42fc-9f4f-95d64b325062" containerName="util" Dec 09 15:45:31 crc kubenswrapper[4894]: I1209 15:45:31.112955 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="caf14a3e-1621-42fc-9f4f-95d64b325062" containerName="extract" Dec 09 15:45:31 crc kubenswrapper[4894]: I1209 15:45:31.113368 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-d7b6w" Dec 09 15:45:31 crc kubenswrapper[4894]: I1209 15:45:31.115367 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-dshrx" Dec 09 15:45:31 crc kubenswrapper[4894]: I1209 15:45:31.141412 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7979d445b4-d7b6w"] Dec 09 15:45:31 crc kubenswrapper[4894]: I1209 15:45:31.226452 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6gblr\" (UniqueName: \"kubernetes.io/projected/463649a6-8336-419d-9017-e62cc33699cd-kube-api-access-6gblr\") pod \"openstack-operator-controller-operator-7979d445b4-d7b6w\" (UID: \"463649a6-8336-419d-9017-e62cc33699cd\") " pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-d7b6w" Dec 09 15:45:31 crc kubenswrapper[4894]: I1209 15:45:31.327899 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6gblr\" (UniqueName: \"kubernetes.io/projected/463649a6-8336-419d-9017-e62cc33699cd-kube-api-access-6gblr\") pod \"openstack-operator-controller-operator-7979d445b4-d7b6w\" (UID: \"463649a6-8336-419d-9017-e62cc33699cd\") " pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-d7b6w" Dec 09 15:45:31 crc kubenswrapper[4894]: I1209 15:45:31.347850 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6gblr\" (UniqueName: \"kubernetes.io/projected/463649a6-8336-419d-9017-e62cc33699cd-kube-api-access-6gblr\") pod \"openstack-operator-controller-operator-7979d445b4-d7b6w\" (UID: \"463649a6-8336-419d-9017-e62cc33699cd\") " pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-d7b6w" Dec 09 15:45:31 crc kubenswrapper[4894]: I1209 15:45:31.434363 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-d7b6w" Dec 09 15:45:31 crc kubenswrapper[4894]: I1209 15:45:31.872185 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7979d445b4-d7b6w"] Dec 09 15:45:32 crc kubenswrapper[4894]: I1209 15:45:32.824220 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-d7b6w" event={"ID":"463649a6-8336-419d-9017-e62cc33699cd","Type":"ContainerStarted","Data":"58b2aead28ef3ca128094eee4b51d92fc4402f0c01024dbbc54a90851344229b"} Dec 09 15:45:36 crc kubenswrapper[4894]: I1209 15:45:36.850389 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-d7b6w" event={"ID":"463649a6-8336-419d-9017-e62cc33699cd","Type":"ContainerStarted","Data":"cef03d77ca55f37dc7b9f83ff3c40435d4574a4b032a1f9ef8295c9e58edf8b0"} Dec 09 15:45:36 crc kubenswrapper[4894]: I1209 15:45:36.851201 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-d7b6w" Dec 09 15:45:36 crc kubenswrapper[4894]: I1209 15:45:36.882777 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-d7b6w" podStartSLOduration=1.1034265 podStartE2EDuration="5.882761828s" podCreationTimestamp="2025-12-09 15:45:31 +0000 UTC" firstStartedPulling="2025-12-09 15:45:31.892121307 +0000 UTC m=+826.211331976" lastFinishedPulling="2025-12-09 15:45:36.671456635 +0000 UTC m=+830.990667304" observedRunningTime="2025-12-09 15:45:36.880071656 +0000 UTC m=+831.199282335" watchObservedRunningTime="2025-12-09 15:45:36.882761828 +0000 UTC m=+831.201972487" Dec 09 15:45:41 crc kubenswrapper[4894]: I1209 15:45:41.439982 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-d7b6w" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.103325 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-xbxkp"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.104967 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-xbxkp" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.108557 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-76dw2" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.113869 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-xbxkp"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.139255 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-rjhh2"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.140415 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-rjhh2" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.144217 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-n555r" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.149303 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-w47lh"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.150515 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-w47lh" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.155957 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-smq9r" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.181616 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-w47lh"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.193950 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-nv5mv"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.194864 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-nv5mv" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.197702 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-6r8cd" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.246722 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-rjhh2"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.251336 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-nv5mv"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.267898 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hlrvt\" (UniqueName: \"kubernetes.io/projected/cfbfae03-83af-4dd6-b04b-6e094638d96e-kube-api-access-hlrvt\") pod \"cinder-operator-controller-manager-6c677c69b-rjhh2\" (UID: \"cfbfae03-83af-4dd6-b04b-6e094638d96e\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-rjhh2" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.268019 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mstmf\" (UniqueName: \"kubernetes.io/projected/52b39bc2-f99f-4865-8fb2-d4377ef35e88-kube-api-access-mstmf\") pod \"barbican-operator-controller-manager-7d9dfd778-xbxkp\" (UID: \"52b39bc2-f99f-4865-8fb2-d4377ef35e88\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-xbxkp" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.268049 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p6wz8\" (UniqueName: \"kubernetes.io/projected/7b9d28bb-988b-4b71-99bb-df62f57debc8-kube-api-access-p6wz8\") pod \"designate-operator-controller-manager-697fb699cf-w47lh\" (UID: \"7b9d28bb-988b-4b71-99bb-df62f57debc8\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-w47lh" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.279267 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-v2jx9"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.293542 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-v2jx9" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.312687 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-vtx6v" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.320522 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6d2gw"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.343537 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6d2gw" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.358231 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-s5dvb" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.367332 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-v2jx9"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.372263 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hlrvt\" (UniqueName: \"kubernetes.io/projected/cfbfae03-83af-4dd6-b04b-6e094638d96e-kube-api-access-hlrvt\") pod \"cinder-operator-controller-manager-6c677c69b-rjhh2\" (UID: \"cfbfae03-83af-4dd6-b04b-6e094638d96e\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-rjhh2" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.372353 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mstmf\" (UniqueName: \"kubernetes.io/projected/52b39bc2-f99f-4865-8fb2-d4377ef35e88-kube-api-access-mstmf\") pod \"barbican-operator-controller-manager-7d9dfd778-xbxkp\" (UID: \"52b39bc2-f99f-4865-8fb2-d4377ef35e88\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-xbxkp" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.372372 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p6wz8\" (UniqueName: \"kubernetes.io/projected/7b9d28bb-988b-4b71-99bb-df62f57debc8-kube-api-access-p6wz8\") pod \"designate-operator-controller-manager-697fb699cf-w47lh\" (UID: \"7b9d28bb-988b-4b71-99bb-df62f57debc8\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-w47lh" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.372410 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-thts8\" (UniqueName: \"kubernetes.io/projected/811e1bfc-20a8-4268-9bda-88dd4e1b33ad-kube-api-access-thts8\") pod \"glance-operator-controller-manager-5697bb5779-nv5mv\" (UID: \"811e1bfc-20a8-4268-9bda-88dd4e1b33ad\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-nv5mv" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.372432 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph2d9\" (UniqueName: \"kubernetes.io/projected/e2902997-82a3-4e80-b50b-6fbc3f290e61-kube-api-access-ph2d9\") pod \"heat-operator-controller-manager-5f64f6f8bb-v2jx9\" (UID: \"e2902997-82a3-4e80-b50b-6fbc3f290e61\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-v2jx9" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.375739 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6d2gw"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.393279 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-mxngk"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.395087 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxngk" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.398508 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-4qzlb"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.400184 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-4qzlb" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.406983 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-5kgk4"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.408540 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5kgk4" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.412374 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-29bb5" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.424484 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-mxngk"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.425179 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-twr66" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.425377 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-jjb4q" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.425809 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.440690 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-4qzlb"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.456259 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-5kgk4"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.458040 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hlrvt\" (UniqueName: \"kubernetes.io/projected/cfbfae03-83af-4dd6-b04b-6e094638d96e-kube-api-access-hlrvt\") pod \"cinder-operator-controller-manager-6c677c69b-rjhh2\" (UID: \"cfbfae03-83af-4dd6-b04b-6e094638d96e\") " pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-rjhh2" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.461329 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p6wz8\" (UniqueName: \"kubernetes.io/projected/7b9d28bb-988b-4b71-99bb-df62f57debc8-kube-api-access-p6wz8\") pod \"designate-operator-controller-manager-697fb699cf-w47lh\" (UID: \"7b9d28bb-988b-4b71-99bb-df62f57debc8\") " pod="openstack-operators/designate-operator-controller-manager-697fb699cf-w47lh" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.468192 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-5ptth"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.469399 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-5ptth" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.476948 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9wtvr"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.478348 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9wtvr" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.479782 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mxngk\" (UID: \"622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxngk" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.479914 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-86wb7\" (UniqueName: \"kubernetes.io/projected/622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7-kube-api-access-86wb7\") pod \"infra-operator-controller-manager-78d48bff9d-mxngk\" (UID: \"622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxngk" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.479964 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2sh4j\" (UniqueName: \"kubernetes.io/projected/7dd34c7e-4139-4dce-873d-32ed62e502ae-kube-api-access-2sh4j\") pod \"ironic-operator-controller-manager-967d97867-4qzlb\" (UID: \"7dd34c7e-4139-4dce-873d-32ed62e502ae\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-4qzlb" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.479993 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-thts8\" (UniqueName: \"kubernetes.io/projected/811e1bfc-20a8-4268-9bda-88dd4e1b33ad-kube-api-access-thts8\") pod \"glance-operator-controller-manager-5697bb5779-nv5mv\" (UID: \"811e1bfc-20a8-4268-9bda-88dd4e1b33ad\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-nv5mv" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.480018 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph2d9\" (UniqueName: \"kubernetes.io/projected/e2902997-82a3-4e80-b50b-6fbc3f290e61-kube-api-access-ph2d9\") pod \"heat-operator-controller-manager-5f64f6f8bb-v2jx9\" (UID: \"e2902997-82a3-4e80-b50b-6fbc3f290e61\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-v2jx9" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.480047 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nsltc\" (UniqueName: \"kubernetes.io/projected/63f9a2a8-9f28-4ee7-a564-9c449f15917b-kube-api-access-nsltc\") pod \"horizon-operator-controller-manager-68c6d99b8f-6d2gw\" (UID: \"63f9a2a8-9f28-4ee7-a564-9c449f15917b\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6d2gw" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.483057 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-rvq7h" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.487894 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-5ptth"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.488151 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-rjhh2" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.492508 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-hflhq" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.502928 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-w47lh" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.503515 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9wtvr"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.509703 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b966c"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.510739 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b966c" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.519933 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-thts8\" (UniqueName: \"kubernetes.io/projected/811e1bfc-20a8-4268-9bda-88dd4e1b33ad-kube-api-access-thts8\") pod \"glance-operator-controller-manager-5697bb5779-nv5mv\" (UID: \"811e1bfc-20a8-4268-9bda-88dd4e1b33ad\") " pod="openstack-operators/glance-operator-controller-manager-5697bb5779-nv5mv" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.523181 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-nv5mv" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.540731 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-b5nj9" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.553963 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b966c"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.581579 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nsltc\" (UniqueName: \"kubernetes.io/projected/63f9a2a8-9f28-4ee7-a564-9c449f15917b-kube-api-access-nsltc\") pod \"horizon-operator-controller-manager-68c6d99b8f-6d2gw\" (UID: \"63f9a2a8-9f28-4ee7-a564-9c449f15917b\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6d2gw" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.581661 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mxngk\" (UID: \"622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxngk" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.581706 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vk62d\" (UniqueName: \"kubernetes.io/projected/d85fd4c5-d08d-4bda-b8b9-aad2165e930c-kube-api-access-vk62d\") pod \"manila-operator-controller-manager-5b5fd79c9c-5ptth\" (UID: \"d85fd4c5-d08d-4bda-b8b9-aad2165e930c\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-5ptth" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.581744 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-556qr\" (UniqueName: \"kubernetes.io/projected/4aa1a97a-d5a6-42e1-9db7-94ca1cad21ff-kube-api-access-556qr\") pod \"keystone-operator-controller-manager-7765d96ddf-5kgk4\" (UID: \"4aa1a97a-d5a6-42e1-9db7-94ca1cad21ff\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5kgk4" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.581779 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2sqv\" (UniqueName: \"kubernetes.io/projected/1fb9fefc-7b00-4116-abc7-6f8bdf847f82-kube-api-access-h2sqv\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-b966c\" (UID: \"1fb9fefc-7b00-4116-abc7-6f8bdf847f82\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b966c" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.581809 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-86wb7\" (UniqueName: \"kubernetes.io/projected/622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7-kube-api-access-86wb7\") pod \"infra-operator-controller-manager-78d48bff9d-mxngk\" (UID: \"622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxngk" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.581830 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pxkm8\" (UniqueName: \"kubernetes.io/projected/ab257a96-149b-4abf-b4b1-8d1beb8e873d-kube-api-access-pxkm8\") pod \"mariadb-operator-controller-manager-79c8c4686c-9wtvr\" (UID: \"ab257a96-149b-4abf-b4b1-8d1beb8e873d\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9wtvr" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.581859 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2sh4j\" (UniqueName: \"kubernetes.io/projected/7dd34c7e-4139-4dce-873d-32ed62e502ae-kube-api-access-2sh4j\") pod \"ironic-operator-controller-manager-967d97867-4qzlb\" (UID: \"7dd34c7e-4139-4dce-873d-32ed62e502ae\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-4qzlb" Dec 09 15:46:01 crc kubenswrapper[4894]: E1209 15:46:01.582320 4894 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 09 15:46:01 crc kubenswrapper[4894]: E1209 15:46:01.582369 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7-cert podName:622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7 nodeName:}" failed. No retries permitted until 2025-12-09 15:46:02.082353302 +0000 UTC m=+856.401563971 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7-cert") pod "infra-operator-controller-manager-78d48bff9d-mxngk" (UID: "622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7") : secret "infra-operator-webhook-server-cert" not found Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.608337 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph2d9\" (UniqueName: \"kubernetes.io/projected/e2902997-82a3-4e80-b50b-6fbc3f290e61-kube-api-access-ph2d9\") pod \"heat-operator-controller-manager-5f64f6f8bb-v2jx9\" (UID: \"e2902997-82a3-4e80-b50b-6fbc3f290e61\") " pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-v2jx9" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.608798 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mstmf\" (UniqueName: \"kubernetes.io/projected/52b39bc2-f99f-4865-8fb2-d4377ef35e88-kube-api-access-mstmf\") pod \"barbican-operator-controller-manager-7d9dfd778-xbxkp\" (UID: \"52b39bc2-f99f-4865-8fb2-d4377ef35e88\") " pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-xbxkp" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.611465 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2sh4j\" (UniqueName: \"kubernetes.io/projected/7dd34c7e-4139-4dce-873d-32ed62e502ae-kube-api-access-2sh4j\") pod \"ironic-operator-controller-manager-967d97867-4qzlb\" (UID: \"7dd34c7e-4139-4dce-873d-32ed62e502ae\") " pod="openstack-operators/ironic-operator-controller-manager-967d97867-4qzlb" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.623717 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-7qv58"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.624793 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-7qv58" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.636839 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-dl6j8" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.642705 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-mhx7f"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.643814 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-mhx7f" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.645197 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-86wb7\" (UniqueName: \"kubernetes.io/projected/622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7-kube-api-access-86wb7\") pod \"infra-operator-controller-manager-78d48bff9d-mxngk\" (UID: \"622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxngk" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.650379 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-rrh74" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.650668 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-v2jx9" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.657239 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nsltc\" (UniqueName: \"kubernetes.io/projected/63f9a2a8-9f28-4ee7-a564-9c449f15917b-kube-api-access-nsltc\") pod \"horizon-operator-controller-manager-68c6d99b8f-6d2gw\" (UID: \"63f9a2a8-9f28-4ee7-a564-9c449f15917b\") " pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6d2gw" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.672076 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6d2gw" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.690712 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-556qr\" (UniqueName: \"kubernetes.io/projected/4aa1a97a-d5a6-42e1-9db7-94ca1cad21ff-kube-api-access-556qr\") pod \"keystone-operator-controller-manager-7765d96ddf-5kgk4\" (UID: \"4aa1a97a-d5a6-42e1-9db7-94ca1cad21ff\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5kgk4" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.690759 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2sqv\" (UniqueName: \"kubernetes.io/projected/1fb9fefc-7b00-4116-abc7-6f8bdf847f82-kube-api-access-h2sqv\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-b966c\" (UID: \"1fb9fefc-7b00-4116-abc7-6f8bdf847f82\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b966c" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.690794 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pxkm8\" (UniqueName: \"kubernetes.io/projected/ab257a96-149b-4abf-b4b1-8d1beb8e873d-kube-api-access-pxkm8\") pod \"mariadb-operator-controller-manager-79c8c4686c-9wtvr\" (UID: \"ab257a96-149b-4abf-b4b1-8d1beb8e873d\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9wtvr" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.690862 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vk62d\" (UniqueName: \"kubernetes.io/projected/d85fd4c5-d08d-4bda-b8b9-aad2165e930c-kube-api-access-vk62d\") pod \"manila-operator-controller-manager-5b5fd79c9c-5ptth\" (UID: \"d85fd4c5-d08d-4bda-b8b9-aad2165e930c\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-5ptth" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.715212 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-mhx7f"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.719674 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vk62d\" (UniqueName: \"kubernetes.io/projected/d85fd4c5-d08d-4bda-b8b9-aad2165e930c-kube-api-access-vk62d\") pod \"manila-operator-controller-manager-5b5fd79c9c-5ptth\" (UID: \"d85fd4c5-d08d-4bda-b8b9-aad2165e930c\") " pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-5ptth" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.726500 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pxkm8\" (UniqueName: \"kubernetes.io/projected/ab257a96-149b-4abf-b4b1-8d1beb8e873d-kube-api-access-pxkm8\") pod \"mariadb-operator-controller-manager-79c8c4686c-9wtvr\" (UID: \"ab257a96-149b-4abf-b4b1-8d1beb8e873d\") " pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9wtvr" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.735013 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9wtvr" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.735375 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-7qv58"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.763823 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-tdcqr"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.764395 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-556qr\" (UniqueName: \"kubernetes.io/projected/4aa1a97a-d5a6-42e1-9db7-94ca1cad21ff-kube-api-access-556qr\") pod \"keystone-operator-controller-manager-7765d96ddf-5kgk4\" (UID: \"4aa1a97a-d5a6-42e1-9db7-94ca1cad21ff\") " pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5kgk4" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.765578 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-xbxkp" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.766184 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-tdcqr" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.785691 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-h6flh" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.786974 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2sqv\" (UniqueName: \"kubernetes.io/projected/1fb9fefc-7b00-4116-abc7-6f8bdf847f82-kube-api-access-h2sqv\") pod \"neutron-operator-controller-manager-5fdfd5b6b5-b966c\" (UID: \"1fb9fefc-7b00-4116-abc7-6f8bdf847f82\") " pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b966c" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.787057 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.788023 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.794404 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-tdcqr"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.795068 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzrfd\" (UniqueName: \"kubernetes.io/projected/cfcfd730-cad4-4774-91fe-7c3d7692f011-kube-api-access-lzrfd\") pod \"octavia-operator-controller-manager-998648c74-mhx7f\" (UID: \"cfcfd730-cad4-4774-91fe-7c3d7692f011\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-mhx7f" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.795123 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k9pxd\" (UniqueName: \"kubernetes.io/projected/e177f09b-5191-4e61-af98-91aa6bf25fc3-kube-api-access-k9pxd\") pod \"nova-operator-controller-manager-697bc559fc-7qv58\" (UID: \"e177f09b-5191-4e61-af98-91aa6bf25fc3\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-7qv58" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.795915 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.804015 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-wxcz2"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.805089 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-wxcz2" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.805835 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-9d65c" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.808400 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-tshxl" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.815579 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b966c" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.824693 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-967d97867-4qzlb" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.903094 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-wpj5r"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.930359 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wpj5r" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.934359 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zhvlj\" (UniqueName: \"kubernetes.io/projected/66650287-6d2b-412e-8b94-fcfedd5a6194-kube-api-access-zhvlj\") pod \"swift-operator-controller-manager-9d58d64bc-wxcz2\" (UID: \"66650287-6d2b-412e-8b94-fcfedd5a6194\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-wxcz2" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.934452 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvvxn\" (UniqueName: \"kubernetes.io/projected/4ca57622-56aa-4588-a1bd-f28972322b58-kube-api-access-kvvxn\") pod \"ovn-operator-controller-manager-b6456fdb6-tdcqr\" (UID: \"4ca57622-56aa-4588-a1bd-f28972322b58\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-tdcqr" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.934598 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9557c24e-8469-46ec-9bab-04e337606beb-cert\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd667tpw\" (UID: \"9557c24e-8469-46ec-9bab-04e337606beb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.934629 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xql6n\" (UniqueName: \"kubernetes.io/projected/9557c24e-8469-46ec-9bab-04e337606beb-kube-api-access-xql6n\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd667tpw\" (UID: \"9557c24e-8469-46ec-9bab-04e337606beb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.934849 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzrfd\" (UniqueName: \"kubernetes.io/projected/cfcfd730-cad4-4774-91fe-7c3d7692f011-kube-api-access-lzrfd\") pod \"octavia-operator-controller-manager-998648c74-mhx7f\" (UID: \"cfcfd730-cad4-4774-91fe-7c3d7692f011\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-mhx7f" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.937761 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k9pxd\" (UniqueName: \"kubernetes.io/projected/e177f09b-5191-4e61-af98-91aa6bf25fc3-kube-api-access-k9pxd\") pod \"nova-operator-controller-manager-697bc559fc-7qv58\" (UID: \"e177f09b-5191-4e61-af98-91aa6bf25fc3\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-7qv58" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.941027 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-tsxc7" Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.981926 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw"] Dec 09 15:46:01 crc kubenswrapper[4894]: I1209 15:46:01.982536 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5kgk4" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.003563 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k9pxd\" (UniqueName: \"kubernetes.io/projected/e177f09b-5191-4e61-af98-91aa6bf25fc3-kube-api-access-k9pxd\") pod \"nova-operator-controller-manager-697bc559fc-7qv58\" (UID: \"e177f09b-5191-4e61-af98-91aa6bf25fc3\") " pod="openstack-operators/nova-operator-controller-manager-697bc559fc-7qv58" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.007144 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzrfd\" (UniqueName: \"kubernetes.io/projected/cfcfd730-cad4-4774-91fe-7c3d7692f011-kube-api-access-lzrfd\") pod \"octavia-operator-controller-manager-998648c74-mhx7f\" (UID: \"cfcfd730-cad4-4774-91fe-7c3d7692f011\") " pod="openstack-operators/octavia-operator-controller-manager-998648c74-mhx7f" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.016312 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-5ptth" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.046206 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-frzr9"] Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.047668 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-frzr9" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.048395 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9557c24e-8469-46ec-9bab-04e337606beb-cert\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd667tpw\" (UID: \"9557c24e-8469-46ec-9bab-04e337606beb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.048463 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xql6n\" (UniqueName: \"kubernetes.io/projected/9557c24e-8469-46ec-9bab-04e337606beb-kube-api-access-xql6n\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd667tpw\" (UID: \"9557c24e-8469-46ec-9bab-04e337606beb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.048515 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5n8x2\" (UniqueName: \"kubernetes.io/projected/faf8439c-1fc7-4316-8b58-31c5a892258b-kube-api-access-5n8x2\") pod \"placement-operator-controller-manager-78f8948974-wpj5r\" (UID: \"faf8439c-1fc7-4316-8b58-31c5a892258b\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-wpj5r" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.048657 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zhvlj\" (UniqueName: \"kubernetes.io/projected/66650287-6d2b-412e-8b94-fcfedd5a6194-kube-api-access-zhvlj\") pod \"swift-operator-controller-manager-9d58d64bc-wxcz2\" (UID: \"66650287-6d2b-412e-8b94-fcfedd5a6194\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-wxcz2" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.048704 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvvxn\" (UniqueName: \"kubernetes.io/projected/4ca57622-56aa-4588-a1bd-f28972322b58-kube-api-access-kvvxn\") pod \"ovn-operator-controller-manager-b6456fdb6-tdcqr\" (UID: \"4ca57622-56aa-4588-a1bd-f28972322b58\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-tdcqr" Dec 09 15:46:02 crc kubenswrapper[4894]: E1209 15:46:02.048998 4894 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 15:46:02 crc kubenswrapper[4894]: E1209 15:46:02.049055 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9557c24e-8469-46ec-9bab-04e337606beb-cert podName:9557c24e-8469-46ec-9bab-04e337606beb nodeName:}" failed. No retries permitted until 2025-12-09 15:46:02.549039466 +0000 UTC m=+856.868250135 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9557c24e-8469-46ec-9bab-04e337606beb-cert") pod "openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" (UID: "9557c24e-8469-46ec-9bab-04e337606beb") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.050322 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-wxcz2"] Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.054064 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-qpcw9" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.057700 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-wpj5r"] Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.077809 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-frzr9"] Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.079037 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvvxn\" (UniqueName: \"kubernetes.io/projected/4ca57622-56aa-4588-a1bd-f28972322b58-kube-api-access-kvvxn\") pod \"ovn-operator-controller-manager-b6456fdb6-tdcqr\" (UID: \"4ca57622-56aa-4588-a1bd-f28972322b58\") " pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-tdcqr" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.082834 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zhvlj\" (UniqueName: \"kubernetes.io/projected/66650287-6d2b-412e-8b94-fcfedd5a6194-kube-api-access-zhvlj\") pod \"swift-operator-controller-manager-9d58d64bc-wxcz2\" (UID: \"66650287-6d2b-412e-8b94-fcfedd5a6194\") " pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-wxcz2" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.088719 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8"] Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.090016 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.091737 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-vfgck" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.092509 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xql6n\" (UniqueName: \"kubernetes.io/projected/9557c24e-8469-46ec-9bab-04e337606beb-kube-api-access-xql6n\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd667tpw\" (UID: \"9557c24e-8469-46ec-9bab-04e337606beb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.100937 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-c9v7t"] Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.102727 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-c9v7t" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.109495 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-k2ndn" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.149621 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5n8x2\" (UniqueName: \"kubernetes.io/projected/faf8439c-1fc7-4316-8b58-31c5a892258b-kube-api-access-5n8x2\") pod \"placement-operator-controller-manager-78f8948974-wpj5r\" (UID: \"faf8439c-1fc7-4316-8b58-31c5a892258b\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-wpj5r" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.149781 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8bnbh\" (UniqueName: \"kubernetes.io/projected/7f99fc36-cb37-456c-bb67-5b4f27d00b80-kube-api-access-8bnbh\") pod \"telemetry-operator-controller-manager-58d5ff84df-frzr9\" (UID: \"7f99fc36-cb37-456c-bb67-5b4f27d00b80\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-frzr9" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.149839 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mxngk\" (UID: \"622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxngk" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.149886 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-wxcz2" Dec 09 15:46:02 crc kubenswrapper[4894]: E1209 15:46:02.149975 4894 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 09 15:46:02 crc kubenswrapper[4894]: E1209 15:46:02.150024 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7-cert podName:622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7 nodeName:}" failed. No retries permitted until 2025-12-09 15:46:03.150009078 +0000 UTC m=+857.469219757 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7-cert") pod "infra-operator-controller-manager-78d48bff9d-mxngk" (UID: "622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7") : secret "infra-operator-webhook-server-cert" not found Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.167358 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8"] Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.168775 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-c9v7t"] Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.174657 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-7qv58" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.184056 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5n8x2\" (UniqueName: \"kubernetes.io/projected/faf8439c-1fc7-4316-8b58-31c5a892258b-kube-api-access-5n8x2\") pod \"placement-operator-controller-manager-78f8948974-wpj5r\" (UID: \"faf8439c-1fc7-4316-8b58-31c5a892258b\") " pod="openstack-operators/placement-operator-controller-manager-78f8948974-wpj5r" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.193487 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wpj5r" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.236149 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-998648c74-mhx7f" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.238425 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f"] Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.247098 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.248792 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f"] Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.251452 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8bnbh\" (UniqueName: \"kubernetes.io/projected/7f99fc36-cb37-456c-bb67-5b4f27d00b80-kube-api-access-8bnbh\") pod \"telemetry-operator-controller-manager-58d5ff84df-frzr9\" (UID: \"7f99fc36-cb37-456c-bb67-5b4f27d00b80\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-frzr9" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.252773 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6g2r\" (UniqueName: \"kubernetes.io/projected/29ac0821-c99d-4b02-8ea7-9985e535305c-kube-api-access-t6g2r\") pod \"watcher-operator-controller-manager-667bd8d554-c9v7t\" (UID: \"29ac0821-c99d-4b02-8ea7-9985e535305c\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-c9v7t" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.253066 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6crfq\" (UniqueName: \"kubernetes.io/projected/93107d61-a76c-4ccf-90c3-4277ac94abf0-kube-api-access-6crfq\") pod \"test-operator-controller-manager-5854674fcc-qtsl8\" (UID: \"93107d61-a76c-4ccf-90c3-4277ac94abf0\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.265163 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.265354 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-mpk4p" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.265461 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"metrics-server-cert" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.277280 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8bnbh\" (UniqueName: \"kubernetes.io/projected/7f99fc36-cb37-456c-bb67-5b4f27d00b80-kube-api-access-8bnbh\") pod \"telemetry-operator-controller-manager-58d5ff84df-frzr9\" (UID: \"7f99fc36-cb37-456c-bb67-5b4f27d00b80\") " pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-frzr9" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.311254 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-tdcqr" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.311978 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6fdzd"] Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.313953 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6fdzd" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.316135 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-r4xr7" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.343097 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6fdzd"] Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.353803 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6crfq\" (UniqueName: \"kubernetes.io/projected/93107d61-a76c-4ccf-90c3-4277ac94abf0-kube-api-access-6crfq\") pod \"test-operator-controller-manager-5854674fcc-qtsl8\" (UID: \"93107d61-a76c-4ccf-90c3-4277ac94abf0\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.353862 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-metrics-certs\") pod \"openstack-operator-controller-manager-668858c49-xgc5f\" (UID: \"cc2f225f-a9cd-493a-bb6b-6d8adb308080\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.353893 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4khd\" (UniqueName: \"kubernetes.io/projected/cc2f225f-a9cd-493a-bb6b-6d8adb308080-kube-api-access-b4khd\") pod \"openstack-operator-controller-manager-668858c49-xgc5f\" (UID: \"cc2f225f-a9cd-493a-bb6b-6d8adb308080\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.353912 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6g2r\" (UniqueName: \"kubernetes.io/projected/29ac0821-c99d-4b02-8ea7-9985e535305c-kube-api-access-t6g2r\") pod \"watcher-operator-controller-manager-667bd8d554-c9v7t\" (UID: \"29ac0821-c99d-4b02-8ea7-9985e535305c\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-c9v7t" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.353932 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-webhook-certs\") pod \"openstack-operator-controller-manager-668858c49-xgc5f\" (UID: \"cc2f225f-a9cd-493a-bb6b-6d8adb308080\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.374883 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6g2r\" (UniqueName: \"kubernetes.io/projected/29ac0821-c99d-4b02-8ea7-9985e535305c-kube-api-access-t6g2r\") pod \"watcher-operator-controller-manager-667bd8d554-c9v7t\" (UID: \"29ac0821-c99d-4b02-8ea7-9985e535305c\") " pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-c9v7t" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.376104 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6crfq\" (UniqueName: \"kubernetes.io/projected/93107d61-a76c-4ccf-90c3-4277ac94abf0-kube-api-access-6crfq\") pod \"test-operator-controller-manager-5854674fcc-qtsl8\" (UID: \"93107d61-a76c-4ccf-90c3-4277ac94abf0\") " pod="openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.456917 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-metrics-certs\") pod \"openstack-operator-controller-manager-668858c49-xgc5f\" (UID: \"cc2f225f-a9cd-493a-bb6b-6d8adb308080\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.456983 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4khd\" (UniqueName: \"kubernetes.io/projected/cc2f225f-a9cd-493a-bb6b-6d8adb308080-kube-api-access-b4khd\") pod \"openstack-operator-controller-manager-668858c49-xgc5f\" (UID: \"cc2f225f-a9cd-493a-bb6b-6d8adb308080\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.457010 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-webhook-certs\") pod \"openstack-operator-controller-manager-668858c49-xgc5f\" (UID: \"cc2f225f-a9cd-493a-bb6b-6d8adb308080\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.457047 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kg66r\" (UniqueName: \"kubernetes.io/projected/7700fe8f-1bda-4f92-81ff-04463e0e4fdf-kube-api-access-kg66r\") pod \"rabbitmq-cluster-operator-manager-668c99d594-6fdzd\" (UID: \"7700fe8f-1bda-4f92-81ff-04463e0e4fdf\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6fdzd" Dec 09 15:46:02 crc kubenswrapper[4894]: E1209 15:46:02.457193 4894 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 09 15:46:02 crc kubenswrapper[4894]: E1209 15:46:02.457246 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-metrics-certs podName:cc2f225f-a9cd-493a-bb6b-6d8adb308080 nodeName:}" failed. No retries permitted until 2025-12-09 15:46:02.957227171 +0000 UTC m=+857.276437840 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-metrics-certs") pod "openstack-operator-controller-manager-668858c49-xgc5f" (UID: "cc2f225f-a9cd-493a-bb6b-6d8adb308080") : secret "metrics-server-cert" not found Dec 09 15:46:02 crc kubenswrapper[4894]: E1209 15:46:02.459541 4894 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 15:46:02 crc kubenswrapper[4894]: E1209 15:46:02.459582 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-webhook-certs podName:cc2f225f-a9cd-493a-bb6b-6d8adb308080 nodeName:}" failed. No retries permitted until 2025-12-09 15:46:02.959570155 +0000 UTC m=+857.278780824 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-webhook-certs") pod "openstack-operator-controller-manager-668858c49-xgc5f" (UID: "cc2f225f-a9cd-493a-bb6b-6d8adb308080") : secret "webhook-server-cert" not found Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.484403 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4khd\" (UniqueName: \"kubernetes.io/projected/cc2f225f-a9cd-493a-bb6b-6d8adb308080-kube-api-access-b4khd\") pod \"openstack-operator-controller-manager-668858c49-xgc5f\" (UID: \"cc2f225f-a9cd-493a-bb6b-6d8adb308080\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.517459 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-6c677c69b-rjhh2"] Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.524692 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-frzr9" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.558083 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9557c24e-8469-46ec-9bab-04e337606beb-cert\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd667tpw\" (UID: \"9557c24e-8469-46ec-9bab-04e337606beb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" Dec 09 15:46:02 crc kubenswrapper[4894]: E1209 15:46:02.558217 4894 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.558249 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kg66r\" (UniqueName: \"kubernetes.io/projected/7700fe8f-1bda-4f92-81ff-04463e0e4fdf-kube-api-access-kg66r\") pod \"rabbitmq-cluster-operator-manager-668c99d594-6fdzd\" (UID: \"7700fe8f-1bda-4f92-81ff-04463e0e4fdf\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6fdzd" Dec 09 15:46:02 crc kubenswrapper[4894]: E1209 15:46:02.558286 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9557c24e-8469-46ec-9bab-04e337606beb-cert podName:9557c24e-8469-46ec-9bab-04e337606beb nodeName:}" failed. No retries permitted until 2025-12-09 15:46:03.558269166 +0000 UTC m=+857.877479835 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9557c24e-8469-46ec-9bab-04e337606beb-cert") pod "openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" (UID: "9557c24e-8469-46ec-9bab-04e337606beb") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.567935 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.580590 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kg66r\" (UniqueName: \"kubernetes.io/projected/7700fe8f-1bda-4f92-81ff-04463e0e4fdf-kube-api-access-kg66r\") pod \"rabbitmq-cluster-operator-manager-668c99d594-6fdzd\" (UID: \"7700fe8f-1bda-4f92-81ff-04463e0e4fdf\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6fdzd" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.589999 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-c9v7t" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.656148 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-5697bb5779-nv5mv"] Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.667116 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6fdzd" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.667248 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-697fb699cf-w47lh"] Dec 09 15:46:02 crc kubenswrapper[4894]: W1209 15:46:02.721505 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod811e1bfc_20a8_4268_9bda_88dd4e1b33ad.slice/crio-0c30c39647ad8168aa567b5182d7125ee845505d7ff9868a14c627baf72ba074 WatchSource:0}: Error finding container 0c30c39647ad8168aa567b5182d7125ee845505d7ff9868a14c627baf72ba074: Status 404 returned error can't find the container with id 0c30c39647ad8168aa567b5182d7125ee845505d7ff9868a14c627baf72ba074 Dec 09 15:46:02 crc kubenswrapper[4894]: W1209 15:46:02.741706 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7b9d28bb_988b_4b71_99bb_df62f57debc8.slice/crio-ed5d694e3489b4e4aa482c27ae54a3e8ca774e27dcf2a2b6844bce48e4507a0d WatchSource:0}: Error finding container ed5d694e3489b4e4aa482c27ae54a3e8ca774e27dcf2a2b6844bce48e4507a0d: Status 404 returned error can't find the container with id ed5d694e3489b4e4aa482c27ae54a3e8ca774e27dcf2a2b6844bce48e4507a0d Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.832910 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6d2gw"] Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.848704 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b966c"] Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.869015 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-5f64f6f8bb-v2jx9"] Dec 09 15:46:02 crc kubenswrapper[4894]: W1209 15:46:02.892987 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode2902997_82a3_4e80_b50b_6fbc3f290e61.slice/crio-5644376ce22d35089d348338d0d9eb2c36e1d899c6b4c9da89755e28d2cea6ff WatchSource:0}: Error finding container 5644376ce22d35089d348338d0d9eb2c36e1d899c6b4c9da89755e28d2cea6ff: Status 404 returned error can't find the container with id 5644376ce22d35089d348338d0d9eb2c36e1d899c6b4c9da89755e28d2cea6ff Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.966097 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-metrics-certs\") pod \"openstack-operator-controller-manager-668858c49-xgc5f\" (UID: \"cc2f225f-a9cd-493a-bb6b-6d8adb308080\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.966549 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-webhook-certs\") pod \"openstack-operator-controller-manager-668858c49-xgc5f\" (UID: \"cc2f225f-a9cd-493a-bb6b-6d8adb308080\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:02 crc kubenswrapper[4894]: E1209 15:46:02.966346 4894 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 09 15:46:02 crc kubenswrapper[4894]: E1209 15:46:02.966926 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-metrics-certs podName:cc2f225f-a9cd-493a-bb6b-6d8adb308080 nodeName:}" failed. No retries permitted until 2025-12-09 15:46:03.966909994 +0000 UTC m=+858.286120663 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-metrics-certs") pod "openstack-operator-controller-manager-668858c49-xgc5f" (UID: "cc2f225f-a9cd-493a-bb6b-6d8adb308080") : secret "metrics-server-cert" not found Dec 09 15:46:02 crc kubenswrapper[4894]: E1209 15:46:02.966767 4894 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 15:46:02 crc kubenswrapper[4894]: E1209 15:46:02.967431 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-webhook-certs podName:cc2f225f-a9cd-493a-bb6b-6d8adb308080 nodeName:}" failed. No retries permitted until 2025-12-09 15:46:03.967421118 +0000 UTC m=+858.286631787 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-webhook-certs") pod "openstack-operator-controller-manager-668858c49-xgc5f" (UID: "cc2f225f-a9cd-493a-bb6b-6d8adb308080") : secret "webhook-server-cert" not found Dec 09 15:46:02 crc kubenswrapper[4894]: I1209 15:46:02.997486 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9wtvr"] Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.016310 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-7d9dfd778-xbxkp"] Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.046749 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-7765d96ddf-5kgk4"] Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.052424 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-967d97867-4qzlb"] Dec 09 15:46:03 crc kubenswrapper[4894]: W1209 15:46:03.053963 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod4aa1a97a_d5a6_42e1_9db7_94ca1cad21ff.slice/crio-177cfb0fd426f52c2f826bb7e79f33b6d97274b62dda74c65e290afbc35dbcaa WatchSource:0}: Error finding container 177cfb0fd426f52c2f826bb7e79f33b6d97274b62dda74c65e290afbc35dbcaa: Status 404 returned error can't find the container with id 177cfb0fd426f52c2f826bb7e79f33b6d97274b62dda74c65e290afbc35dbcaa Dec 09 15:46:03 crc kubenswrapper[4894]: W1209 15:46:03.065960 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7dd34c7e_4139_4dce_873d_32ed62e502ae.slice/crio-b9a72e25b489d03fc43bbf6f5b6cc9070eeeddc66d2ca41214b13a6f5eaa0ab0 WatchSource:0}: Error finding container b9a72e25b489d03fc43bbf6f5b6cc9070eeeddc66d2ca41214b13a6f5eaa0ab0: Status 404 returned error can't find the container with id b9a72e25b489d03fc43bbf6f5b6cc9070eeeddc66d2ca41214b13a6f5eaa0ab0 Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.076744 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-nv5mv" event={"ID":"811e1bfc-20a8-4268-9bda-88dd4e1b33ad","Type":"ContainerStarted","Data":"0c30c39647ad8168aa567b5182d7125ee845505d7ff9868a14c627baf72ba074"} Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.077808 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-w47lh" event={"ID":"7b9d28bb-988b-4b71-99bb-df62f57debc8","Type":"ContainerStarted","Data":"ed5d694e3489b4e4aa482c27ae54a3e8ca774e27dcf2a2b6844bce48e4507a0d"} Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.079386 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-xbxkp" event={"ID":"52b39bc2-f99f-4865-8fb2-d4377ef35e88","Type":"ContainerStarted","Data":"0588bbdacfe42d2c4481d82b94c0eec0a025af08fb608752a6810b38eaca9dca"} Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.081893 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-v2jx9" event={"ID":"e2902997-82a3-4e80-b50b-6fbc3f290e61","Type":"ContainerStarted","Data":"5644376ce22d35089d348338d0d9eb2c36e1d899c6b4c9da89755e28d2cea6ff"} Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.082983 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9wtvr" event={"ID":"ab257a96-149b-4abf-b4b1-8d1beb8e873d","Type":"ContainerStarted","Data":"7623d70dbe1d141e2fe5666b5bc3abc52fb30cd8410179cef0ea9ade33b97471"} Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.083815 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b966c" event={"ID":"1fb9fefc-7b00-4116-abc7-6f8bdf847f82","Type":"ContainerStarted","Data":"07ca696c540402c1f27651c127e84bd16eebb6a524883c7d370c1ae9c71f7e48"} Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.085854 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-rjhh2" event={"ID":"cfbfae03-83af-4dd6-b04b-6e094638d96e","Type":"ContainerStarted","Data":"2283027f5e47682a84412e600d403cd3e2a0656fad9c7c4499376ee9279387e4"} Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.086968 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6d2gw" event={"ID":"63f9a2a8-9f28-4ee7-a564-9c449f15917b","Type":"ContainerStarted","Data":"7a6e0c4b247402a03dccaf64db42f2a9401e2dd995ad14d9e01264e681dee570"} Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.175427 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mxngk\" (UID: \"622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxngk" Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.175680 4894 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.175725 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7-cert podName:622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7 nodeName:}" failed. No retries permitted until 2025-12-09 15:46:05.175711214 +0000 UTC m=+859.494921883 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7-cert") pod "infra-operator-controller-manager-78d48bff9d-mxngk" (UID: "622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7") : secret "infra-operator-webhook-server-cert" not found Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.176061 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-998648c74-mhx7f"] Dec 09 15:46:03 crc kubenswrapper[4894]: W1209 15:46:03.178366 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcfcfd730_cad4_4774_91fe_7c3d7692f011.slice/crio-eeded9f8efaaef64edc15f13c9a3075204b62ee7dac76bc76a16f615b6789a67 WatchSource:0}: Error finding container eeded9f8efaaef64edc15f13c9a3075204b62ee7dac76bc76a16f615b6789a67: Status 404 returned error can't find the container with id eeded9f8efaaef64edc15f13c9a3075204b62ee7dac76bc76a16f615b6789a67 Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.183543 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-5b5fd79c9c-5ptth"] Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.196765 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-b6456fdb6-tdcqr"] Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.200597 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-9d58d64bc-wxcz2"] Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.201792 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zhvlj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-wxcz2_openstack-operators(66650287-6d2b-412e-8b94-fcfedd5a6194): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.202933 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vk62d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5b5fd79c9c-5ptth_openstack-operators(d85fd4c5-d08d-4bda-b8b9-aad2165e930c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.204680 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-zhvlj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-9d58d64bc-wxcz2_openstack-operators(66650287-6d2b-412e-8b94-fcfedd5a6194): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.205895 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-wxcz2" podUID="66650287-6d2b-412e-8b94-fcfedd5a6194" Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.206486 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-vk62d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod manila-operator-controller-manager-5b5fd79c9c-5ptth_openstack-operators(d85fd4c5-d08d-4bda-b8b9-aad2165e930c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.208422 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-5ptth" podUID="d85fd4c5-d08d-4bda-b8b9-aad2165e930c" Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.307776 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-58d5ff84df-frzr9"] Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.314377 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-78f8948974-wpj5r"] Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.314952 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k9pxd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-7qv58_openstack-operators(e177f09b-5191-4e61-af98-91aa6bf25fc3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 15:46:03 crc kubenswrapper[4894]: W1209 15:46:03.315826 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfaf8439c_1fc7_4316_8b58_31c5a892258b.slice/crio-a20340ddf5044cb1ca26d4b0a4ff9d259f9c511492bf2409a75cd47a5915931d WatchSource:0}: Error finding container a20340ddf5044cb1ca26d4b0a4ff9d259f9c511492bf2409a75cd47a5915931d: Status 404 returned error can't find the container with id a20340ddf5044cb1ca26d4b0a4ff9d259f9c511492bf2409a75cd47a5915931d Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.320046 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-k9pxd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-697bc559fc-7qv58_openstack-operators(e177f09b-5191-4e61-af98-91aa6bf25fc3): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.321501 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-7qv58" podUID="e177f09b-5191-4e61-af98-91aa6bf25fc3" Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.322461 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5n8x2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-wpj5r_openstack-operators(faf8439c-1fc7-4316-8b58-31c5a892258b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.322624 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8bnbh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-frzr9_openstack-operators(7f99fc36-cb37-456c-bb67-5b4f27d00b80): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.322733 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-697bc559fc-7qv58"] Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.325221 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-5n8x2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-78f8948974-wpj5r_openstack-operators(faf8439c-1fc7-4316-8b58-31c5a892258b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.325316 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-8bnbh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-58d5ff84df-frzr9_openstack-operators(7f99fc36-cb37-456c-bb67-5b4f27d00b80): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.326601 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-frzr9" podUID="7f99fc36-cb37-456c-bb67-5b4f27d00b80" Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.326696 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wpj5r" podUID="faf8439c-1fc7-4316-8b58-31c5a892258b" Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.327947 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6crfq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-qtsl8_openstack-operators(93107d61-a76c-4ccf-90c3-4277ac94abf0): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.328207 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6fdzd"] Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.333122 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6crfq,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-5854674fcc-qtsl8_openstack-operators(93107d61-a76c-4ccf-90c3-4277ac94abf0): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.333325 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t6g2r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-667bd8d554-c9v7t_openstack-operators(29ac0821-c99d-4b02-8ea7-9985e535305c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.334843 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8" podUID="93107d61-a76c-4ccf-90c3-4277ac94abf0" Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.335579 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:kube-rbac-proxy,Image:quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0,Command:[],Args:[--secure-listen-address=0.0.0.0:8443 --upstream=http://127.0.0.1:8080/ --logtostderr=true --v=0],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:https,HostPort:0,ContainerPort:8443,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{134217728 0} {} BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-t6g2r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-667bd8d554-c9v7t_openstack-operators(29ac0821-c99d-4b02-8ea7-9985e535305c): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.336948 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ErrImagePull: \"pull QPS exceeded\"]" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-c9v7t" podUID="29ac0821-c99d-4b02-8ea7-9985e535305c" Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.342737 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8"] Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.347345 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-667bd8d554-c9v7t"] Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.583319 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9557c24e-8469-46ec-9bab-04e337606beb-cert\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd667tpw\" (UID: \"9557c24e-8469-46ec-9bab-04e337606beb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.583540 4894 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.583975 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9557c24e-8469-46ec-9bab-04e337606beb-cert podName:9557c24e-8469-46ec-9bab-04e337606beb nodeName:}" failed. No retries permitted until 2025-12-09 15:46:05.583948382 +0000 UTC m=+859.903159051 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9557c24e-8469-46ec-9bab-04e337606beb-cert") pod "openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" (UID: "9557c24e-8469-46ec-9bab-04e337606beb") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.990243 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-metrics-certs\") pod \"openstack-operator-controller-manager-668858c49-xgc5f\" (UID: \"cc2f225f-a9cd-493a-bb6b-6d8adb308080\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:03 crc kubenswrapper[4894]: I1209 15:46:03.990308 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-webhook-certs\") pod \"openstack-operator-controller-manager-668858c49-xgc5f\" (UID: \"cc2f225f-a9cd-493a-bb6b-6d8adb308080\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.990437 4894 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.990497 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-metrics-certs podName:cc2f225f-a9cd-493a-bb6b-6d8adb308080 nodeName:}" failed. No retries permitted until 2025-12-09 15:46:05.990479212 +0000 UTC m=+860.309689881 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-metrics-certs") pod "openstack-operator-controller-manager-668858c49-xgc5f" (UID: "cc2f225f-a9cd-493a-bb6b-6d8adb308080") : secret "metrics-server-cert" not found Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.990588 4894 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 15:46:03 crc kubenswrapper[4894]: E1209 15:46:03.990734 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-webhook-certs podName:cc2f225f-a9cd-493a-bb6b-6d8adb308080 nodeName:}" failed. No retries permitted until 2025-12-09 15:46:05.990710758 +0000 UTC m=+860.309921437 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-webhook-certs") pod "openstack-operator-controller-manager-668858c49-xgc5f" (UID: "cc2f225f-a9cd-493a-bb6b-6d8adb308080") : secret "webhook-server-cert" not found Dec 09 15:46:04 crc kubenswrapper[4894]: E1209 15:46:04.119396 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-5ptth" podUID="d85fd4c5-d08d-4bda-b8b9-aad2165e930c" Dec 09 15:46:04 crc kubenswrapper[4894]: E1209 15:46:04.120401 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8" podUID="93107d61-a76c-4ccf-90c3-4277ac94abf0" Dec 09 15:46:04 crc kubenswrapper[4894]: I1209 15:46:04.120973 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6fdzd" event={"ID":"7700fe8f-1bda-4f92-81ff-04463e0e4fdf","Type":"ContainerStarted","Data":"9bc4bfe0bb631c53a558dc8fe72229b8f2c41beef315500a8448cb6451ca1073"} Dec 09 15:46:04 crc kubenswrapper[4894]: I1209 15:46:04.121007 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-5ptth" event={"ID":"d85fd4c5-d08d-4bda-b8b9-aad2165e930c","Type":"ContainerStarted","Data":"85a21042ba17aa0c65784f333514cfbb7908d49a83a2e2d38d39ade3ffabc696"} Dec 09 15:46:04 crc kubenswrapper[4894]: I1209 15:46:04.121030 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8" event={"ID":"93107d61-a76c-4ccf-90c3-4277ac94abf0","Type":"ContainerStarted","Data":"561407eb4fbb557daf9edd331bec86db91ee822f86ca8ac7e8a1041ccf7fbaae"} Dec 09 15:46:04 crc kubenswrapper[4894]: I1209 15:46:04.121043 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-mhx7f" event={"ID":"cfcfd730-cad4-4774-91fe-7c3d7692f011","Type":"ContainerStarted","Data":"eeded9f8efaaef64edc15f13c9a3075204b62ee7dac76bc76a16f615b6789a67"} Dec 09 15:46:04 crc kubenswrapper[4894]: I1209 15:46:04.122502 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-tdcqr" event={"ID":"4ca57622-56aa-4588-a1bd-f28972322b58","Type":"ContainerStarted","Data":"6a9d226076a262880b6df85e2105958d30c1dd7deeba274b75e5b0e5e5283b2d"} Dec 09 15:46:04 crc kubenswrapper[4894]: I1209 15:46:04.125489 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wpj5r" event={"ID":"faf8439c-1fc7-4316-8b58-31c5a892258b","Type":"ContainerStarted","Data":"a20340ddf5044cb1ca26d4b0a4ff9d259f9c511492bf2409a75cd47a5915931d"} Dec 09 15:46:04 crc kubenswrapper[4894]: I1209 15:46:04.138751 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-7qv58" event={"ID":"e177f09b-5191-4e61-af98-91aa6bf25fc3","Type":"ContainerStarted","Data":"ee6d2da8f6d492e3df4573c720c8baf04c127b0747053c7a375773211148b75f"} Dec 09 15:46:04 crc kubenswrapper[4894]: E1209 15:46:04.139882 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wpj5r" podUID="faf8439c-1fc7-4316-8b58-31c5a892258b" Dec 09 15:46:04 crc kubenswrapper[4894]: I1209 15:46:04.143023 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-4qzlb" event={"ID":"7dd34c7e-4139-4dce-873d-32ed62e502ae","Type":"ContainerStarted","Data":"b9a72e25b489d03fc43bbf6f5b6cc9070eeeddc66d2ca41214b13a6f5eaa0ab0"} Dec 09 15:46:04 crc kubenswrapper[4894]: E1209 15:46:04.144361 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-7qv58" podUID="e177f09b-5191-4e61-af98-91aa6bf25fc3" Dec 09 15:46:04 crc kubenswrapper[4894]: I1209 15:46:04.144680 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-wxcz2" event={"ID":"66650287-6d2b-412e-8b94-fcfedd5a6194","Type":"ContainerStarted","Data":"84a87652fffe6e1c39fc3c7b1307f260d511de5a0cef1adc51df341c4774b6bb"} Dec 09 15:46:04 crc kubenswrapper[4894]: I1209 15:46:04.148716 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5kgk4" event={"ID":"4aa1a97a-d5a6-42e1-9db7-94ca1cad21ff","Type":"ContainerStarted","Data":"177cfb0fd426f52c2f826bb7e79f33b6d97274b62dda74c65e290afbc35dbcaa"} Dec 09 15:46:04 crc kubenswrapper[4894]: I1209 15:46:04.150836 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-frzr9" event={"ID":"7f99fc36-cb37-456c-bb67-5b4f27d00b80","Type":"ContainerStarted","Data":"ccc72689a07b84a128d6bf4991044187e9571645fde9a0f1fc564b54ec00bd7d"} Dec 09 15:46:04 crc kubenswrapper[4894]: E1209 15:46:04.156082 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-frzr9" podUID="7f99fc36-cb37-456c-bb67-5b4f27d00b80" Dec 09 15:46:04 crc kubenswrapper[4894]: E1209 15:46:04.156183 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-wxcz2" podUID="66650287-6d2b-412e-8b94-fcfedd5a6194" Dec 09 15:46:04 crc kubenswrapper[4894]: I1209 15:46:04.162694 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-c9v7t" event={"ID":"29ac0821-c99d-4b02-8ea7-9985e535305c","Type":"ContainerStarted","Data":"ff4f6ac5f2a69936ef6ecebaf359823878e0aa24eac1b6f17bc86f31c8775e41"} Dec 09 15:46:04 crc kubenswrapper[4894]: E1209 15:46:04.169376 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-c9v7t" podUID="29ac0821-c99d-4b02-8ea7-9985e535305c" Dec 09 15:46:04 crc kubenswrapper[4894]: I1209 15:46:04.854263 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-thsb5"] Dec 09 15:46:04 crc kubenswrapper[4894]: I1209 15:46:04.866727 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-thsb5" Dec 09 15:46:04 crc kubenswrapper[4894]: I1209 15:46:04.882626 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-thsb5"] Dec 09 15:46:04 crc kubenswrapper[4894]: I1209 15:46:04.931483 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd8e92cc-f08a-41f2-a439-03fe7c0bc000-utilities\") pod \"redhat-operators-thsb5\" (UID: \"cd8e92cc-f08a-41f2-a439-03fe7c0bc000\") " pod="openshift-marketplace/redhat-operators-thsb5" Dec 09 15:46:04 crc kubenswrapper[4894]: I1209 15:46:04.931607 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfbqn\" (UniqueName: \"kubernetes.io/projected/cd8e92cc-f08a-41f2-a439-03fe7c0bc000-kube-api-access-tfbqn\") pod \"redhat-operators-thsb5\" (UID: \"cd8e92cc-f08a-41f2-a439-03fe7c0bc000\") " pod="openshift-marketplace/redhat-operators-thsb5" Dec 09 15:46:04 crc kubenswrapper[4894]: I1209 15:46:04.931659 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd8e92cc-f08a-41f2-a439-03fe7c0bc000-catalog-content\") pod \"redhat-operators-thsb5\" (UID: \"cd8e92cc-f08a-41f2-a439-03fe7c0bc000\") " pod="openshift-marketplace/redhat-operators-thsb5" Dec 09 15:46:05 crc kubenswrapper[4894]: I1209 15:46:05.032942 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfbqn\" (UniqueName: \"kubernetes.io/projected/cd8e92cc-f08a-41f2-a439-03fe7c0bc000-kube-api-access-tfbqn\") pod \"redhat-operators-thsb5\" (UID: \"cd8e92cc-f08a-41f2-a439-03fe7c0bc000\") " pod="openshift-marketplace/redhat-operators-thsb5" Dec 09 15:46:05 crc kubenswrapper[4894]: I1209 15:46:05.032989 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd8e92cc-f08a-41f2-a439-03fe7c0bc000-catalog-content\") pod \"redhat-operators-thsb5\" (UID: \"cd8e92cc-f08a-41f2-a439-03fe7c0bc000\") " pod="openshift-marketplace/redhat-operators-thsb5" Dec 09 15:46:05 crc kubenswrapper[4894]: I1209 15:46:05.033066 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd8e92cc-f08a-41f2-a439-03fe7c0bc000-utilities\") pod \"redhat-operators-thsb5\" (UID: \"cd8e92cc-f08a-41f2-a439-03fe7c0bc000\") " pod="openshift-marketplace/redhat-operators-thsb5" Dec 09 15:46:05 crc kubenswrapper[4894]: I1209 15:46:05.033499 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd8e92cc-f08a-41f2-a439-03fe7c0bc000-utilities\") pod \"redhat-operators-thsb5\" (UID: \"cd8e92cc-f08a-41f2-a439-03fe7c0bc000\") " pod="openshift-marketplace/redhat-operators-thsb5" Dec 09 15:46:05 crc kubenswrapper[4894]: I1209 15:46:05.033796 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd8e92cc-f08a-41f2-a439-03fe7c0bc000-catalog-content\") pod \"redhat-operators-thsb5\" (UID: \"cd8e92cc-f08a-41f2-a439-03fe7c0bc000\") " pod="openshift-marketplace/redhat-operators-thsb5" Dec 09 15:46:05 crc kubenswrapper[4894]: I1209 15:46:05.068221 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfbqn\" (UniqueName: \"kubernetes.io/projected/cd8e92cc-f08a-41f2-a439-03fe7c0bc000-kube-api-access-tfbqn\") pod \"redhat-operators-thsb5\" (UID: \"cd8e92cc-f08a-41f2-a439-03fe7c0bc000\") " pod="openshift-marketplace/redhat-operators-thsb5" Dec 09 15:46:05 crc kubenswrapper[4894]: E1209 15:46:05.186423 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8" podUID="93107d61-a76c-4ccf-90c3-4277ac94abf0" Dec 09 15:46:05 crc kubenswrapper[4894]: E1209 15:46:05.186687 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:6b3e0302608a2e70f9b5ae9167f6fbf59264f226d9db99d48f70466ab2f216b8\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-c9v7t" podUID="29ac0821-c99d-4b02-8ea7-9985e535305c" Dec 09 15:46:05 crc kubenswrapper[4894]: E1209 15:46:05.186762 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:779f0cee6024d0fb8f259b036fe790e62aa5a3b0431ea9bf15a6e7d02e2e5670\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-7qv58" podUID="e177f09b-5191-4e61-af98-91aa6bf25fc3" Dec 09 15:46:05 crc kubenswrapper[4894]: E1209 15:46:05.186850 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d29650b006da97eb9178fcc58f2eb9fead8c2b414fac18f86a3c3a1507488c4f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wpj5r" podUID="faf8439c-1fc7-4316-8b58-31c5a892258b" Dec 09 15:46:05 crc kubenswrapper[4894]: E1209 15:46:05.186940 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:f27e732ec1faee765461bf137d9be81278b2fa39675019a73622755e1e610b6f\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-frzr9" podUID="7f99fc36-cb37-456c-bb67-5b4f27d00b80" Dec 09 15:46:05 crc kubenswrapper[4894]: E1209 15:46:05.188509 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:3aa109bb973253ae9dcf339b9b65abbd1176cdb4be672c93e538a5f113816991\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-wxcz2" podUID="66650287-6d2b-412e-8b94-fcfedd5a6194" Dec 09 15:46:05 crc kubenswrapper[4894]: E1209 15:46:05.191834 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/manila-operator@sha256:44126f9c6b1d2bf752ddf989e20a4fc4cc1c07723d4fcb78465ccb2f55da6b3a\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-5ptth" podUID="d85fd4c5-d08d-4bda-b8b9-aad2165e930c" Dec 09 15:46:05 crc kubenswrapper[4894]: I1209 15:46:05.235540 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mxngk\" (UID: \"622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxngk" Dec 09 15:46:05 crc kubenswrapper[4894]: E1209 15:46:05.235682 4894 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 09 15:46:05 crc kubenswrapper[4894]: E1209 15:46:05.235722 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7-cert podName:622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7 nodeName:}" failed. No retries permitted until 2025-12-09 15:46:09.235707708 +0000 UTC m=+863.554918377 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7-cert") pod "infra-operator-controller-manager-78d48bff9d-mxngk" (UID: "622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7") : secret "infra-operator-webhook-server-cert" not found Dec 09 15:46:05 crc kubenswrapper[4894]: I1209 15:46:05.239476 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-thsb5" Dec 09 15:46:05 crc kubenswrapper[4894]: I1209 15:46:05.641693 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9557c24e-8469-46ec-9bab-04e337606beb-cert\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd667tpw\" (UID: \"9557c24e-8469-46ec-9bab-04e337606beb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" Dec 09 15:46:05 crc kubenswrapper[4894]: E1209 15:46:05.641849 4894 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 15:46:05 crc kubenswrapper[4894]: E1209 15:46:05.641917 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9557c24e-8469-46ec-9bab-04e337606beb-cert podName:9557c24e-8469-46ec-9bab-04e337606beb nodeName:}" failed. No retries permitted until 2025-12-09 15:46:09.64189859 +0000 UTC m=+863.961109259 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9557c24e-8469-46ec-9bab-04e337606beb-cert") pod "openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" (UID: "9557c24e-8469-46ec-9bab-04e337606beb") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 15:46:06 crc kubenswrapper[4894]: I1209 15:46:06.052240 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-metrics-certs\") pod \"openstack-operator-controller-manager-668858c49-xgc5f\" (UID: \"cc2f225f-a9cd-493a-bb6b-6d8adb308080\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:06 crc kubenswrapper[4894]: I1209 15:46:06.052290 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-webhook-certs\") pod \"openstack-operator-controller-manager-668858c49-xgc5f\" (UID: \"cc2f225f-a9cd-493a-bb6b-6d8adb308080\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:06 crc kubenswrapper[4894]: E1209 15:46:06.052418 4894 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 15:46:06 crc kubenswrapper[4894]: E1209 15:46:06.052468 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-webhook-certs podName:cc2f225f-a9cd-493a-bb6b-6d8adb308080 nodeName:}" failed. No retries permitted until 2025-12-09 15:46:10.05245282 +0000 UTC m=+864.371663489 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-webhook-certs") pod "openstack-operator-controller-manager-668858c49-xgc5f" (UID: "cc2f225f-a9cd-493a-bb6b-6d8adb308080") : secret "webhook-server-cert" not found Dec 09 15:46:06 crc kubenswrapper[4894]: E1209 15:46:06.052809 4894 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 09 15:46:06 crc kubenswrapper[4894]: E1209 15:46:06.052861 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-metrics-certs podName:cc2f225f-a9cd-493a-bb6b-6d8adb308080 nodeName:}" failed. No retries permitted until 2025-12-09 15:46:10.0528513 +0000 UTC m=+864.372061959 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-metrics-certs") pod "openstack-operator-controller-manager-668858c49-xgc5f" (UID: "cc2f225f-a9cd-493a-bb6b-6d8adb308080") : secret "metrics-server-cert" not found Dec 09 15:46:06 crc kubenswrapper[4894]: E1209 15:46:06.198196 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:101b3e007d8c9f2e183262d7712f986ad51256448099069bc14f1ea5f997ab94\\\"\", failed to \"StartContainer\" for \"kube-rbac-proxy\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/kube-rbac-proxy:v0.16.0\\\"\"]" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8" podUID="93107d61-a76c-4ccf-90c3-4277ac94abf0" Dec 09 15:46:08 crc kubenswrapper[4894]: I1209 15:46:08.126521 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-nkmzh"] Dec 09 15:46:08 crc kubenswrapper[4894]: I1209 15:46:08.128723 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nkmzh" Dec 09 15:46:08 crc kubenswrapper[4894]: I1209 15:46:08.136734 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nkmzh"] Dec 09 15:46:08 crc kubenswrapper[4894]: I1209 15:46:08.184977 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59332ca3-ed84-4e9b-b965-669f3ecbdd51-utilities\") pod \"community-operators-nkmzh\" (UID: \"59332ca3-ed84-4e9b-b965-669f3ecbdd51\") " pod="openshift-marketplace/community-operators-nkmzh" Dec 09 15:46:08 crc kubenswrapper[4894]: I1209 15:46:08.185071 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59332ca3-ed84-4e9b-b965-669f3ecbdd51-catalog-content\") pod \"community-operators-nkmzh\" (UID: \"59332ca3-ed84-4e9b-b965-669f3ecbdd51\") " pod="openshift-marketplace/community-operators-nkmzh" Dec 09 15:46:08 crc kubenswrapper[4894]: I1209 15:46:08.185097 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wbjn7\" (UniqueName: \"kubernetes.io/projected/59332ca3-ed84-4e9b-b965-669f3ecbdd51-kube-api-access-wbjn7\") pod \"community-operators-nkmzh\" (UID: \"59332ca3-ed84-4e9b-b965-669f3ecbdd51\") " pod="openshift-marketplace/community-operators-nkmzh" Dec 09 15:46:08 crc kubenswrapper[4894]: I1209 15:46:08.286651 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59332ca3-ed84-4e9b-b965-669f3ecbdd51-catalog-content\") pod \"community-operators-nkmzh\" (UID: \"59332ca3-ed84-4e9b-b965-669f3ecbdd51\") " pod="openshift-marketplace/community-operators-nkmzh" Dec 09 15:46:08 crc kubenswrapper[4894]: I1209 15:46:08.286710 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wbjn7\" (UniqueName: \"kubernetes.io/projected/59332ca3-ed84-4e9b-b965-669f3ecbdd51-kube-api-access-wbjn7\") pod \"community-operators-nkmzh\" (UID: \"59332ca3-ed84-4e9b-b965-669f3ecbdd51\") " pod="openshift-marketplace/community-operators-nkmzh" Dec 09 15:46:08 crc kubenswrapper[4894]: I1209 15:46:08.286820 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59332ca3-ed84-4e9b-b965-669f3ecbdd51-utilities\") pod \"community-operators-nkmzh\" (UID: \"59332ca3-ed84-4e9b-b965-669f3ecbdd51\") " pod="openshift-marketplace/community-operators-nkmzh" Dec 09 15:46:08 crc kubenswrapper[4894]: I1209 15:46:08.287197 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59332ca3-ed84-4e9b-b965-669f3ecbdd51-catalog-content\") pod \"community-operators-nkmzh\" (UID: \"59332ca3-ed84-4e9b-b965-669f3ecbdd51\") " pod="openshift-marketplace/community-operators-nkmzh" Dec 09 15:46:08 crc kubenswrapper[4894]: I1209 15:46:08.287262 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59332ca3-ed84-4e9b-b965-669f3ecbdd51-utilities\") pod \"community-operators-nkmzh\" (UID: \"59332ca3-ed84-4e9b-b965-669f3ecbdd51\") " pod="openshift-marketplace/community-operators-nkmzh" Dec 09 15:46:08 crc kubenswrapper[4894]: I1209 15:46:08.308212 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wbjn7\" (UniqueName: \"kubernetes.io/projected/59332ca3-ed84-4e9b-b965-669f3ecbdd51-kube-api-access-wbjn7\") pod \"community-operators-nkmzh\" (UID: \"59332ca3-ed84-4e9b-b965-669f3ecbdd51\") " pod="openshift-marketplace/community-operators-nkmzh" Dec 09 15:46:08 crc kubenswrapper[4894]: I1209 15:46:08.454019 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nkmzh" Dec 09 15:46:09 crc kubenswrapper[4894]: I1209 15:46:09.300480 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mxngk\" (UID: \"622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxngk" Dec 09 15:46:09 crc kubenswrapper[4894]: E1209 15:46:09.300825 4894 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Dec 09 15:46:09 crc kubenswrapper[4894]: E1209 15:46:09.300875 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7-cert podName:622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7 nodeName:}" failed. No retries permitted until 2025-12-09 15:46:17.300859996 +0000 UTC m=+871.620070665 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7-cert") pod "infra-operator-controller-manager-78d48bff9d-mxngk" (UID: "622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7") : secret "infra-operator-webhook-server-cert" not found Dec 09 15:46:09 crc kubenswrapper[4894]: I1209 15:46:09.707355 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9557c24e-8469-46ec-9bab-04e337606beb-cert\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd667tpw\" (UID: \"9557c24e-8469-46ec-9bab-04e337606beb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" Dec 09 15:46:09 crc kubenswrapper[4894]: E1209 15:46:09.707704 4894 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 15:46:09 crc kubenswrapper[4894]: E1209 15:46:09.708812 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9557c24e-8469-46ec-9bab-04e337606beb-cert podName:9557c24e-8469-46ec-9bab-04e337606beb nodeName:}" failed. No retries permitted until 2025-12-09 15:46:17.708389055 +0000 UTC m=+872.027599764 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9557c24e-8469-46ec-9bab-04e337606beb-cert") pod "openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" (UID: "9557c24e-8469-46ec-9bab-04e337606beb") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 15:46:10 crc kubenswrapper[4894]: I1209 15:46:10.113712 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-metrics-certs\") pod \"openstack-operator-controller-manager-668858c49-xgc5f\" (UID: \"cc2f225f-a9cd-493a-bb6b-6d8adb308080\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:10 crc kubenswrapper[4894]: I1209 15:46:10.114354 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-webhook-certs\") pod \"openstack-operator-controller-manager-668858c49-xgc5f\" (UID: \"cc2f225f-a9cd-493a-bb6b-6d8adb308080\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:10 crc kubenswrapper[4894]: E1209 15:46:10.113933 4894 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 09 15:46:10 crc kubenswrapper[4894]: E1209 15:46:10.114477 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-metrics-certs podName:cc2f225f-a9cd-493a-bb6b-6d8adb308080 nodeName:}" failed. No retries permitted until 2025-12-09 15:46:18.114432853 +0000 UTC m=+872.433643522 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-metrics-certs") pod "openstack-operator-controller-manager-668858c49-xgc5f" (UID: "cc2f225f-a9cd-493a-bb6b-6d8adb308080") : secret "metrics-server-cert" not found Dec 09 15:46:10 crc kubenswrapper[4894]: E1209 15:46:10.114790 4894 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 15:46:10 crc kubenswrapper[4894]: E1209 15:46:10.114944 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-webhook-certs podName:cc2f225f-a9cd-493a-bb6b-6d8adb308080 nodeName:}" failed. No retries permitted until 2025-12-09 15:46:18.114888785 +0000 UTC m=+872.434099494 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-webhook-certs") pod "openstack-operator-controller-manager-668858c49-xgc5f" (UID: "cc2f225f-a9cd-493a-bb6b-6d8adb308080") : secret "webhook-server-cert" not found Dec 09 15:46:12 crc kubenswrapper[4894]: I1209 15:46:12.257527 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:46:12 crc kubenswrapper[4894]: I1209 15:46:12.257592 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:46:15 crc kubenswrapper[4894]: E1209 15:46:15.504965 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027" Dec 09 15:46:15 crc kubenswrapper[4894]: E1209 15:46:15.505705 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/glance-operator@sha256:5370dc4a8e776923eec00bb50cbdb2e390e9dde50be26bdc04a216bd2d6b5027,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-thts8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod glance-operator-controller-manager-5697bb5779-nv5mv_openstack-operators(811e1bfc-20a8-4268-9bda-88dd4e1b33ad): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 15:46:17 crc kubenswrapper[4894]: I1209 15:46:17.339725 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mxngk\" (UID: \"622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxngk" Dec 09 15:46:17 crc kubenswrapper[4894]: I1209 15:46:17.351282 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7-cert\") pod \"infra-operator-controller-manager-78d48bff9d-mxngk\" (UID: \"622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7\") " pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxngk" Dec 09 15:46:17 crc kubenswrapper[4894]: I1209 15:46:17.627056 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxngk" Dec 09 15:46:17 crc kubenswrapper[4894]: I1209 15:46:17.747383 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9557c24e-8469-46ec-9bab-04e337606beb-cert\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd667tpw\" (UID: \"9557c24e-8469-46ec-9bab-04e337606beb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" Dec 09 15:46:17 crc kubenswrapper[4894]: E1209 15:46:17.747585 4894 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 15:46:17 crc kubenswrapper[4894]: E1209 15:46:17.747686 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/9557c24e-8469-46ec-9bab-04e337606beb-cert podName:9557c24e-8469-46ec-9bab-04e337606beb nodeName:}" failed. No retries permitted until 2025-12-09 15:46:33.74766561 +0000 UTC m=+888.066876319 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/9557c24e-8469-46ec-9bab-04e337606beb-cert") pod "openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" (UID: "9557c24e-8469-46ec-9bab-04e337606beb") : secret "openstack-baremetal-operator-webhook-server-cert" not found Dec 09 15:46:18 crc kubenswrapper[4894]: I1209 15:46:18.152941 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-metrics-certs\") pod \"openstack-operator-controller-manager-668858c49-xgc5f\" (UID: \"cc2f225f-a9cd-493a-bb6b-6d8adb308080\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:18 crc kubenswrapper[4894]: I1209 15:46:18.153025 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-webhook-certs\") pod \"openstack-operator-controller-manager-668858c49-xgc5f\" (UID: \"cc2f225f-a9cd-493a-bb6b-6d8adb308080\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:18 crc kubenswrapper[4894]: E1209 15:46:18.153135 4894 secret.go:188] Couldn't get secret openstack-operators/metrics-server-cert: secret "metrics-server-cert" not found Dec 09 15:46:18 crc kubenswrapper[4894]: E1209 15:46:18.153173 4894 secret.go:188] Couldn't get secret openstack-operators/webhook-server-cert: secret "webhook-server-cert" not found Dec 09 15:46:18 crc kubenswrapper[4894]: E1209 15:46:18.153201 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-metrics-certs podName:cc2f225f-a9cd-493a-bb6b-6d8adb308080 nodeName:}" failed. No retries permitted until 2025-12-09 15:46:34.153183214 +0000 UTC m=+888.472393883 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-metrics-certs") pod "openstack-operator-controller-manager-668858c49-xgc5f" (UID: "cc2f225f-a9cd-493a-bb6b-6d8adb308080") : secret "metrics-server-cert" not found Dec 09 15:46:18 crc kubenswrapper[4894]: E1209 15:46:18.153215 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-webhook-certs podName:cc2f225f-a9cd-493a-bb6b-6d8adb308080 nodeName:}" failed. No retries permitted until 2025-12-09 15:46:34.153208595 +0000 UTC m=+888.472419264 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "webhook-certs" (UniqueName: "kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-webhook-certs") pod "openstack-operator-controller-manager-668858c49-xgc5f" (UID: "cc2f225f-a9cd-493a-bb6b-6d8adb308080") : secret "webhook-server-cert" not found Dec 09 15:46:26 crc kubenswrapper[4894]: E1209 15:46:26.918367 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7" Dec 09 15:46:26 crc kubenswrapper[4894]: E1209 15:46:26.919107 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:72ad6517987f674af0d0ae092cbb874aeae909c8b8b60188099c311762ebc8f7,Command:[/manager],Args:[--leader-elect --health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-556qr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-7765d96ddf-5kgk4_openstack-operators(4aa1a97a-d5a6-42e1-9db7-94ca1cad21ff): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 15:46:27 crc kubenswrapper[4894]: E1209 15:46:27.788474 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2" Dec 09 15:46:27 crc kubenswrapper[4894]: E1209 15:46:27.788630 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},EnvVar{Name:METRICS_CERTS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-kg66r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-668c99d594-6fdzd_openstack-operators(7700fe8f-1bda-4f92-81ff-04463e0e4fdf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 15:46:27 crc kubenswrapper[4894]: E1209 15:46:27.789788 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6fdzd" podUID="7700fe8f-1bda-4f92-81ff-04463e0e4fdf" Dec 09 15:46:28 crc kubenswrapper[4894]: I1209 15:46:28.345794 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-mhx7f" event={"ID":"cfcfd730-cad4-4774-91fe-7c3d7692f011","Type":"ContainerStarted","Data":"d3db0ea6963a65a17df4d1ef1100fb9a625c0d9ae2ee14d48e2f4a07b62940c5"} Dec 09 15:46:28 crc kubenswrapper[4894]: I1209 15:46:28.349258 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-xbxkp" event={"ID":"52b39bc2-f99f-4865-8fb2-d4377ef35e88","Type":"ContainerStarted","Data":"fb3573fbe35db4ef9f861abbf7ee71f981a9e1fe460dc5604714a701baeaaf70"} Dec 09 15:46:28 crc kubenswrapper[4894]: I1209 15:46:28.352467 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6d2gw" event={"ID":"63f9a2a8-9f28-4ee7-a564-9c449f15917b","Type":"ContainerStarted","Data":"e4fe8f04f246f90f12529b7b97a3722e07a9858106de4d9cd8db292b6504dbb3"} Dec 09 15:46:28 crc kubenswrapper[4894]: E1209 15:46:28.354551 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6fdzd" podUID="7700fe8f-1bda-4f92-81ff-04463e0e4fdf" Dec 09 15:46:28 crc kubenswrapper[4894]: I1209 15:46:28.469709 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-78d48bff9d-mxngk"] Dec 09 15:46:28 crc kubenswrapper[4894]: I1209 15:46:28.477032 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-thsb5"] Dec 09 15:46:28 crc kubenswrapper[4894]: I1209 15:46:28.527347 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-nkmzh"] Dec 09 15:46:28 crc kubenswrapper[4894]: W1209 15:46:28.748594 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod622a8fae_c8e4_4ec9_bd3f_edbb1e2a5ed7.slice/crio-68e9ba3f11c9b8683e93860f9191d9d26ebdb3f29dd79a01f444b9d33ea3c413 WatchSource:0}: Error finding container 68e9ba3f11c9b8683e93860f9191d9d26ebdb3f29dd79a01f444b9d33ea3c413: Status 404 returned error can't find the container with id 68e9ba3f11c9b8683e93860f9191d9d26ebdb3f29dd79a01f444b9d33ea3c413 Dec 09 15:46:29 crc kubenswrapper[4894]: I1209 15:46:29.366478 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-rjhh2" event={"ID":"cfbfae03-83af-4dd6-b04b-6e094638d96e","Type":"ContainerStarted","Data":"1000e8c20d3d59804ad5d32ed762d06ff2b252eb8aa559a0b47c467f262097e5"} Dec 09 15:46:29 crc kubenswrapper[4894]: I1209 15:46:29.368983 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-thsb5" event={"ID":"cd8e92cc-f08a-41f2-a439-03fe7c0bc000","Type":"ContainerStarted","Data":"d19e1a38a6a18767e020576bb99489f7498a8feaf99d252d6a03be3715eff920"} Dec 09 15:46:29 crc kubenswrapper[4894]: I1209 15:46:29.370262 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxngk" event={"ID":"622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7","Type":"ContainerStarted","Data":"68e9ba3f11c9b8683e93860f9191d9d26ebdb3f29dd79a01f444b9d33ea3c413"} Dec 09 15:46:29 crc kubenswrapper[4894]: I1209 15:46:29.374721 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b966c" event={"ID":"1fb9fefc-7b00-4116-abc7-6f8bdf847f82","Type":"ContainerStarted","Data":"c7e20517e854079d9cd7056b108a6305f4b067ccf95a3cdbd291df831409ef4f"} Dec 09 15:46:29 crc kubenswrapper[4894]: I1209 15:46:29.377237 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-tdcqr" event={"ID":"4ca57622-56aa-4588-a1bd-f28972322b58","Type":"ContainerStarted","Data":"64d98e799b3157308e58e9c41f4d9cd7d6f39478171738ede16959efacb4f3c3"} Dec 09 15:46:29 crc kubenswrapper[4894]: I1209 15:46:29.379004 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-w47lh" event={"ID":"7b9d28bb-988b-4b71-99bb-df62f57debc8","Type":"ContainerStarted","Data":"35fae6fffc932dc72190645c1c6304cbf977460619e979a4d3ab7484e49e5163"} Dec 09 15:46:31 crc kubenswrapper[4894]: I1209 15:46:31.392404 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nkmzh" event={"ID":"59332ca3-ed84-4e9b-b965-669f3ecbdd51","Type":"ContainerStarted","Data":"3b9b10724e1e536853a73416337b75d496ac25049377f78927aaeae67da39e14"} Dec 09 15:46:33 crc kubenswrapper[4894]: I1209 15:46:33.812808 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9557c24e-8469-46ec-9bab-04e337606beb-cert\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd667tpw\" (UID: \"9557c24e-8469-46ec-9bab-04e337606beb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" Dec 09 15:46:33 crc kubenswrapper[4894]: I1209 15:46:33.819048 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/9557c24e-8469-46ec-9bab-04e337606beb-cert\") pod \"openstack-baremetal-operator-controller-manager-694d6cfbd667tpw\" (UID: \"9557c24e-8469-46ec-9bab-04e337606beb\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" Dec 09 15:46:33 crc kubenswrapper[4894]: I1209 15:46:33.901698 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" Dec 09 15:46:34 crc kubenswrapper[4894]: I1209 15:46:34.218666 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-metrics-certs\") pod \"openstack-operator-controller-manager-668858c49-xgc5f\" (UID: \"cc2f225f-a9cd-493a-bb6b-6d8adb308080\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:34 crc kubenswrapper[4894]: I1209 15:46:34.218729 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-webhook-certs\") pod \"openstack-operator-controller-manager-668858c49-xgc5f\" (UID: \"cc2f225f-a9cd-493a-bb6b-6d8adb308080\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:34 crc kubenswrapper[4894]: I1209 15:46:34.222852 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-webhook-certs\") pod \"openstack-operator-controller-manager-668858c49-xgc5f\" (UID: \"cc2f225f-a9cd-493a-bb6b-6d8adb308080\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:34 crc kubenswrapper[4894]: I1209 15:46:34.236516 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/cc2f225f-a9cd-493a-bb6b-6d8adb308080-metrics-certs\") pod \"openstack-operator-controller-manager-668858c49-xgc5f\" (UID: \"cc2f225f-a9cd-493a-bb6b-6d8adb308080\") " pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:34 crc kubenswrapper[4894]: I1209 15:46:34.413056 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:35 crc kubenswrapper[4894]: I1209 15:46:35.425154 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-4qzlb" event={"ID":"7dd34c7e-4139-4dce-873d-32ed62e502ae","Type":"ContainerStarted","Data":"5f6baf684b9b66718783614d93073aaa690c26b15f5ad483922a50e94fc82b7d"} Dec 09 15:46:35 crc kubenswrapper[4894]: I1209 15:46:35.427975 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9wtvr" event={"ID":"ab257a96-149b-4abf-b4b1-8d1beb8e873d","Type":"ContainerStarted","Data":"879c59a853aedb671d772db256d42e628ecba9b9568e1403c9fab42b8a67be1c"} Dec 09 15:46:36 crc kubenswrapper[4894]: I1209 15:46:36.273838 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f"] Dec 09 15:46:36 crc kubenswrapper[4894]: I1209 15:46:36.435154 4894 generic.go:334] "Generic (PLEG): container finished" podID="59332ca3-ed84-4e9b-b965-669f3ecbdd51" containerID="61bda99fa3c7b7c2fd7a05cdec3621c434d888aee71787e243e1cf4956e64dfc" exitCode=0 Dec 09 15:46:36 crc kubenswrapper[4894]: I1209 15:46:36.435203 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nkmzh" event={"ID":"59332ca3-ed84-4e9b-b965-669f3ecbdd51","Type":"ContainerDied","Data":"61bda99fa3c7b7c2fd7a05cdec3621c434d888aee71787e243e1cf4956e64dfc"} Dec 09 15:46:36 crc kubenswrapper[4894]: I1209 15:46:36.443344 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-v2jx9" event={"ID":"e2902997-82a3-4e80-b50b-6fbc3f290e61","Type":"ContainerStarted","Data":"d739de9e7236477dac2a564795b98a5d38c86bf4c676e43fb984fefad8707084"} Dec 09 15:46:36 crc kubenswrapper[4894]: I1209 15:46:36.445511 4894 generic.go:334] "Generic (PLEG): container finished" podID="cd8e92cc-f08a-41f2-a439-03fe7c0bc000" containerID="3d74115cfd0e8ae4a5fb19fa386ac82eccb6627741cbd6107bcfd721cd6198e4" exitCode=0 Dec 09 15:46:36 crc kubenswrapper[4894]: I1209 15:46:36.445532 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-thsb5" event={"ID":"cd8e92cc-f08a-41f2-a439-03fe7c0bc000","Type":"ContainerDied","Data":"3d74115cfd0e8ae4a5fb19fa386ac82eccb6627741cbd6107bcfd721cd6198e4"} Dec 09 15:46:36 crc kubenswrapper[4894]: I1209 15:46:36.467222 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw"] Dec 09 15:46:36 crc kubenswrapper[4894]: W1209 15:46:36.818843 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9557c24e_8469_46ec_9bab_04e337606beb.slice/crio-953626c6372f2f7f6c0b3c5d06517bd75a31d261eeaae6a891dafeba38aeb3ff WatchSource:0}: Error finding container 953626c6372f2f7f6c0b3c5d06517bd75a31d261eeaae6a891dafeba38aeb3ff: Status 404 returned error can't find the container with id 953626c6372f2f7f6c0b3c5d06517bd75a31d261eeaae6a891dafeba38aeb3ff Dec 09 15:46:37 crc kubenswrapper[4894]: E1209 15:46:37.494916 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-nv5mv" podUID="811e1bfc-20a8-4268-9bda-88dd4e1b33ad" Dec 09 15:46:37 crc kubenswrapper[4894]: I1209 15:46:37.517692 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-5ptth" event={"ID":"d85fd4c5-d08d-4bda-b8b9-aad2165e930c","Type":"ContainerStarted","Data":"40a1abb13f16de7d6d04dcdc53bdb5f71be582dc4909a07cf78e18f89ea669db"} Dec 09 15:46:37 crc kubenswrapper[4894]: I1209 15:46:37.530627 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" event={"ID":"9557c24e-8469-46ec-9bab-04e337606beb","Type":"ContainerStarted","Data":"953626c6372f2f7f6c0b3c5d06517bd75a31d261eeaae6a891dafeba38aeb3ff"} Dec 09 15:46:37 crc kubenswrapper[4894]: I1209 15:46:37.578935 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8" event={"ID":"93107d61-a76c-4ccf-90c3-4277ac94abf0","Type":"ContainerStarted","Data":"3e4d7bf3c4557c742c5380fe554f925744aef6d53f7e1bebe4a47f5e0d3390d2"} Dec 09 15:46:37 crc kubenswrapper[4894]: I1209 15:46:37.588784 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-frzr9" event={"ID":"7f99fc36-cb37-456c-bb67-5b4f27d00b80","Type":"ContainerStarted","Data":"2df1867a64d2d0c3c5b9b8ba1d03583bd0b9e0a2145223c677154074780331b7"} Dec 09 15:46:37 crc kubenswrapper[4894]: I1209 15:46:37.616251 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wpj5r" event={"ID":"faf8439c-1fc7-4316-8b58-31c5a892258b","Type":"ContainerStarted","Data":"ccef46af345b58aed8f3048b959e4c3b3568855d0bbd26807e55e7b9325f9900"} Dec 09 15:46:37 crc kubenswrapper[4894]: I1209 15:46:37.636019 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" event={"ID":"cc2f225f-a9cd-493a-bb6b-6d8adb308080","Type":"ContainerStarted","Data":"15e15d5a875613d80cb132902db9fe88e01be6fdb21df2e891b6ca89b5f8111d"} Dec 09 15:46:37 crc kubenswrapper[4894]: I1209 15:46:37.636057 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" event={"ID":"cc2f225f-a9cd-493a-bb6b-6d8adb308080","Type":"ContainerStarted","Data":"ea36aa64aeda92cabde2b1b62a6402da1e1b0d3a0631662a0fd83675a4e36b41"} Dec 09 15:46:37 crc kubenswrapper[4894]: I1209 15:46:37.636650 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:37 crc kubenswrapper[4894]: I1209 15:46:37.668951 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-c9v7t" event={"ID":"29ac0821-c99d-4b02-8ea7-9985e535305c","Type":"ContainerStarted","Data":"96998369a020a27e3619bb0a4b6ab0ec7f4151b266cd0799af8f4fabf9d8f327"} Dec 09 15:46:37 crc kubenswrapper[4894]: I1209 15:46:37.690239 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-7qv58" event={"ID":"e177f09b-5191-4e61-af98-91aa6bf25fc3","Type":"ContainerStarted","Data":"7e66c476e0b9b280dc98a947e285cb88cff46b2a4d3f2ee2789c796c16ff07f0"} Dec 09 15:46:37 crc kubenswrapper[4894]: I1209 15:46:37.736459 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" podStartSLOduration=36.736426768 podStartE2EDuration="36.736426768s" podCreationTimestamp="2025-12-09 15:46:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:46:37.696115591 +0000 UTC m=+892.015326260" watchObservedRunningTime="2025-12-09 15:46:37.736426768 +0000 UTC m=+892.055637437" Dec 09 15:46:37 crc kubenswrapper[4894]: I1209 15:46:37.738009 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-wxcz2" event={"ID":"66650287-6d2b-412e-8b94-fcfedd5a6194","Type":"ContainerStarted","Data":"416fd92e890c540d18a3832cf60f87023a4495ee8f69eaa8d162e25211860b5b"} Dec 09 15:46:37 crc kubenswrapper[4894]: I1209 15:46:37.738235 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-wxcz2" Dec 09 15:46:37 crc kubenswrapper[4894]: I1209 15:46:37.759044 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxngk" event={"ID":"622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7","Type":"ContainerStarted","Data":"2d997f381b1851cce10db1ea7bde273dee6d239f58b4ebd4666faac3eadfb99d"} Dec 09 15:46:37 crc kubenswrapper[4894]: I1209 15:46:37.798084 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-wxcz2" podStartSLOduration=4.591591114 podStartE2EDuration="36.79806665s" podCreationTimestamp="2025-12-09 15:46:01 +0000 UTC" firstStartedPulling="2025-12-09 15:46:03.201582052 +0000 UTC m=+857.520792721" lastFinishedPulling="2025-12-09 15:46:35.408057588 +0000 UTC m=+889.727268257" observedRunningTime="2025-12-09 15:46:37.775947104 +0000 UTC m=+892.095157773" watchObservedRunningTime="2025-12-09 15:46:37.79806665 +0000 UTC m=+892.117277309" Dec 09 15:46:37 crc kubenswrapper[4894]: E1209 15:46:37.976045 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5kgk4" podUID="4aa1a97a-d5a6-42e1-9db7-94ca1cad21ff" Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.808022 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-rjhh2" event={"ID":"cfbfae03-83af-4dd6-b04b-6e094638d96e","Type":"ContainerStarted","Data":"9b556e0c384043e42c60be91a4de6c589696ddfc437f7d987f1de6211867b207"} Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.808461 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-rjhh2" Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.810988 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-rjhh2" Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.821304 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-nv5mv" event={"ID":"811e1bfc-20a8-4268-9bda-88dd4e1b33ad","Type":"ContainerStarted","Data":"266382d8fb6f73477ef5d4d824724fdf619d3b1aee1c0c87a91ef9e61761584a"} Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.828669 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-xbxkp" event={"ID":"52b39bc2-f99f-4865-8fb2-d4377ef35e88","Type":"ContainerStarted","Data":"5022670bcc79cf27b8382b00ca427696c05cc1d1280f461ad1e7f8721740c770"} Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.829389 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-xbxkp" Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.838237 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-xbxkp" Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.841673 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-6c677c69b-rjhh2" podStartSLOduration=3.358801523 podStartE2EDuration="37.841624877s" podCreationTimestamp="2025-12-09 15:46:01 +0000 UTC" firstStartedPulling="2025-12-09 15:46:02.626026163 +0000 UTC m=+856.945236832" lastFinishedPulling="2025-12-09 15:46:37.108849527 +0000 UTC m=+891.428060186" observedRunningTime="2025-12-09 15:46:38.837781714 +0000 UTC m=+893.156992383" watchObservedRunningTime="2025-12-09 15:46:38.841624877 +0000 UTC m=+893.160835546" Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.859097 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b966c" event={"ID":"1fb9fefc-7b00-4116-abc7-6f8bdf847f82","Type":"ContainerStarted","Data":"ecc312ead51b7d574b6b19d8af8964b36048f257540a9b47e028a8124b66586a"} Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.860809 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b966c" Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.864507 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b966c" Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.869543 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-wxcz2" event={"ID":"66650287-6d2b-412e-8b94-fcfedd5a6194","Type":"ContainerStarted","Data":"e1859d8ed6f29714c153027f0547b1fdd137e6a4ba803b7cea3cab45434dab78"} Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.872053 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-frzr9" event={"ID":"7f99fc36-cb37-456c-bb67-5b4f27d00b80","Type":"ContainerStarted","Data":"537a99e75d7454ace11372221a8505065a7dda1f6198adb0ae2d55b102e706b9"} Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.874855 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-frzr9" Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.883079 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-7d9dfd778-xbxkp" podStartSLOduration=3.572351982 podStartE2EDuration="37.883057555s" podCreationTimestamp="2025-12-09 15:46:01 +0000 UTC" firstStartedPulling="2025-12-09 15:46:03.028237568 +0000 UTC m=+857.347448237" lastFinishedPulling="2025-12-09 15:46:37.338943141 +0000 UTC m=+891.658153810" observedRunningTime="2025-12-09 15:46:38.878985745 +0000 UTC m=+893.198196414" watchObservedRunningTime="2025-12-09 15:46:38.883057555 +0000 UTC m=+893.202268224" Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.894969 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6d2gw" event={"ID":"63f9a2a8-9f28-4ee7-a564-9c449f15917b","Type":"ContainerStarted","Data":"56979b78b87ac0339b84e67182b8e528c0ec0e76355ffde3a8cca2b3f415d01c"} Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.895203 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6d2gw" Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.896986 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6d2gw" Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.914825 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8" event={"ID":"93107d61-a76c-4ccf-90c3-4277ac94abf0","Type":"ContainerStarted","Data":"d5cb7a833eb3803d582a2aece60dd5f9df69dbbbc23ef667e3422c98f7f4d8b3"} Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.915814 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8" Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.930777 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-v2jx9" event={"ID":"e2902997-82a3-4e80-b50b-6fbc3f290e61","Type":"ContainerStarted","Data":"25fb8d6cdfd099f45592841e08349a5391e8b6db4d52cd9395742c6f7b8e02af"} Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.931799 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-v2jx9" Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.939370 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-998648c74-mhx7f" event={"ID":"cfcfd730-cad4-4774-91fe-7c3d7692f011","Type":"ContainerStarted","Data":"f6a4b3ec0d5419a8c0504791f554d61fa763c861ea092c600534a6ec42b55bad"} Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.940443 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-998648c74-mhx7f" Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.948526 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-5fdfd5b6b5-b966c" podStartSLOduration=3.521406598 podStartE2EDuration="37.9485061s" podCreationTimestamp="2025-12-09 15:46:01 +0000 UTC" firstStartedPulling="2025-12-09 15:46:02.856427755 +0000 UTC m=+857.175638424" lastFinishedPulling="2025-12-09 15:46:37.283527257 +0000 UTC m=+891.602737926" observedRunningTime="2025-12-09 15:46:38.943650868 +0000 UTC m=+893.262861547" watchObservedRunningTime="2025-12-09 15:46:38.9485061 +0000 UTC m=+893.267716779" Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.957026 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-998648c74-mhx7f" Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.967681 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-967d97867-4qzlb" event={"ID":"7dd34c7e-4139-4dce-873d-32ed62e502ae","Type":"ContainerStarted","Data":"1355c827b8a160da92e12b061ce5b2315b2b98c1a74f78fb0e992f68810503f8"} Dec 09 15:46:38 crc kubenswrapper[4894]: I1209 15:46:38.969169 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-967d97867-4qzlb" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.008580 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9wtvr" event={"ID":"ab257a96-149b-4abf-b4b1-8d1beb8e873d","Type":"ContainerStarted","Data":"e72165a1251abeefe4d48a1b3115845c424911aaa859e27ffdce28c5e15a0dcc"} Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.008992 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9wtvr" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.016903 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-68c6d99b8f-6d2gw" podStartSLOduration=3.761399839 podStartE2EDuration="38.016888194s" podCreationTimestamp="2025-12-09 15:46:01 +0000 UTC" firstStartedPulling="2025-12-09 15:46:02.858720967 +0000 UTC m=+857.177931636" lastFinishedPulling="2025-12-09 15:46:37.114209322 +0000 UTC m=+891.433419991" observedRunningTime="2025-12-09 15:46:39.013839571 +0000 UTC m=+893.333050250" watchObservedRunningTime="2025-12-09 15:46:39.016888194 +0000 UTC m=+893.336098863" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.027933 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9wtvr" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.035376 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-tdcqr" event={"ID":"4ca57622-56aa-4588-a1bd-f28972322b58","Type":"ContainerStarted","Data":"65374ae34f08d236eee469e742871ac21f27ddbee9870131025d4aaed167006d"} Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.043963 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-tdcqr" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.059963 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-tdcqr" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.069756 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-c9v7t" event={"ID":"29ac0821-c99d-4b02-8ea7-9985e535305c","Type":"ContainerStarted","Data":"1699be1fa503460d11ce899d0c2f705d9c2e059fe3a4bf658a847959a91f7c31"} Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.070724 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-c9v7t" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.077120 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-frzr9" podStartSLOduration=6.664740963 podStartE2EDuration="38.077090587s" podCreationTimestamp="2025-12-09 15:46:01 +0000 UTC" firstStartedPulling="2025-12-09 15:46:03.322540953 +0000 UTC m=+857.641751622" lastFinishedPulling="2025-12-09 15:46:34.734890587 +0000 UTC m=+889.054101246" observedRunningTime="2025-12-09 15:46:39.055838343 +0000 UTC m=+893.375049012" watchObservedRunningTime="2025-12-09 15:46:39.077090587 +0000 UTC m=+893.396301266" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.103136 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nkmzh" event={"ID":"59332ca3-ed84-4e9b-b965-669f3ecbdd51","Type":"ContainerStarted","Data":"ba3f78c7c2e5fa81a4a818c1e0d985756064884f87b381464bcdfd47dcb21f52"} Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.138327 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-b6456fdb6-tdcqr" podStartSLOduration=4.214700581 podStartE2EDuration="38.138289726s" podCreationTimestamp="2025-12-09 15:46:01 +0000 UTC" firstStartedPulling="2025-12-09 15:46:03.201368796 +0000 UTC m=+857.520579465" lastFinishedPulling="2025-12-09 15:46:37.124957941 +0000 UTC m=+891.444168610" observedRunningTime="2025-12-09 15:46:39.106017966 +0000 UTC m=+893.425228635" watchObservedRunningTime="2025-12-09 15:46:39.138289726 +0000 UTC m=+893.457500395" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.155510 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-w47lh" event={"ID":"7b9d28bb-988b-4b71-99bb-df62f57debc8","Type":"ContainerStarted","Data":"2bd98e419a80b6d718408fc43f6e2f4b0665d2f47f3b298799f958b1a2fd1047"} Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.156457 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-w47lh" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.162933 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8" podStartSLOduration=6.080841239 podStartE2EDuration="38.162919561s" podCreationTimestamp="2025-12-09 15:46:01 +0000 UTC" firstStartedPulling="2025-12-09 15:46:03.327824116 +0000 UTC m=+857.647034785" lastFinishedPulling="2025-12-09 15:46:35.409902438 +0000 UTC m=+889.729113107" observedRunningTime="2025-12-09 15:46:39.156028005 +0000 UTC m=+893.475238684" watchObservedRunningTime="2025-12-09 15:46:39.162919561 +0000 UTC m=+893.482130230" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.177026 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-w47lh" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.183885 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxngk" event={"ID":"622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7","Type":"ContainerStarted","Data":"376a0ff3353c70c7de1d90b593f143e1a9a6dd30280f28e86b28e7f17ed79faf"} Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.184587 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxngk" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.200070 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5kgk4" event={"ID":"4aa1a97a-d5a6-42e1-9db7-94ca1cad21ff","Type":"ContainerStarted","Data":"441fc18aaef2c4977926ae55c6669bd46e81d4af75892dfe7bfd5fd09335248c"} Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.212334 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-79c8c4686c-9wtvr" podStartSLOduration=4.106435222 podStartE2EDuration="38.212309812s" podCreationTimestamp="2025-12-09 15:46:01 +0000 UTC" firstStartedPulling="2025-12-09 15:46:03.007954321 +0000 UTC m=+857.327164990" lastFinishedPulling="2025-12-09 15:46:37.113828911 +0000 UTC m=+891.433039580" observedRunningTime="2025-12-09 15:46:39.19813636 +0000 UTC m=+893.517347049" watchObservedRunningTime="2025-12-09 15:46:39.212309812 +0000 UTC m=+893.531520481" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.240846 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-v2jx9" podStartSLOduration=3.593637587 podStartE2EDuration="38.240821292s" podCreationTimestamp="2025-12-09 15:46:01 +0000 UTC" firstStartedPulling="2025-12-09 15:46:02.902889479 +0000 UTC m=+857.222100148" lastFinishedPulling="2025-12-09 15:46:37.550073184 +0000 UTC m=+891.869283853" observedRunningTime="2025-12-09 15:46:39.234207543 +0000 UTC m=+893.553418212" watchObservedRunningTime="2025-12-09 15:46:39.240821292 +0000 UTC m=+893.560031961" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.257628 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-967d97867-4qzlb" podStartSLOduration=4.14046328 podStartE2EDuration="38.257607974s" podCreationTimestamp="2025-12-09 15:46:01 +0000 UTC" firstStartedPulling="2025-12-09 15:46:03.075257186 +0000 UTC m=+857.394467845" lastFinishedPulling="2025-12-09 15:46:37.19240187 +0000 UTC m=+891.511612539" observedRunningTime="2025-12-09 15:46:39.25523882 +0000 UTC m=+893.574449489" watchObservedRunningTime="2025-12-09 15:46:39.257607974 +0000 UTC m=+893.576818633" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.289682 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-998648c74-mhx7f" podStartSLOduration=4.41939395 podStartE2EDuration="38.289663508s" podCreationTimestamp="2025-12-09 15:46:01 +0000 UTC" firstStartedPulling="2025-12-09 15:46:03.18371473 +0000 UTC m=+857.502925389" lastFinishedPulling="2025-12-09 15:46:37.053984278 +0000 UTC m=+891.373194947" observedRunningTime="2025-12-09 15:46:39.287813828 +0000 UTC m=+893.607024497" watchObservedRunningTime="2025-12-09 15:46:39.289663508 +0000 UTC m=+893.608874177" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.376986 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxngk" podStartSLOduration=31.154792971 podStartE2EDuration="38.376969052s" podCreationTimestamp="2025-12-09 15:46:01 +0000 UTC" firstStartedPulling="2025-12-09 15:46:28.769139654 +0000 UTC m=+883.088350323" lastFinishedPulling="2025-12-09 15:46:35.991315735 +0000 UTC m=+890.310526404" observedRunningTime="2025-12-09 15:46:39.376135159 +0000 UTC m=+893.695345828" watchObservedRunningTime="2025-12-09 15:46:39.376969052 +0000 UTC m=+893.696179721" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.396019 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-jxpm6"] Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.399206 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jxpm6" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.404694 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jxpm6"] Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.410491 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-c9v7t" podStartSLOduration=7.007634138 podStartE2EDuration="38.410470726s" podCreationTimestamp="2025-12-09 15:46:01 +0000 UTC" firstStartedPulling="2025-12-09 15:46:03.33318566 +0000 UTC m=+857.652396329" lastFinishedPulling="2025-12-09 15:46:34.736022208 +0000 UTC m=+889.055232917" observedRunningTime="2025-12-09 15:46:39.408756469 +0000 UTC m=+893.727967138" watchObservedRunningTime="2025-12-09 15:46:39.410470726 +0000 UTC m=+893.729681405" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.483308 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-697fb699cf-w47lh" podStartSLOduration=4.055314154 podStartE2EDuration="38.483282359s" podCreationTimestamp="2025-12-09 15:46:01 +0000 UTC" firstStartedPulling="2025-12-09 15:46:02.763848219 +0000 UTC m=+857.083058888" lastFinishedPulling="2025-12-09 15:46:37.191816424 +0000 UTC m=+891.511027093" observedRunningTime="2025-12-09 15:46:39.4729734 +0000 UTC m=+893.792184079" watchObservedRunningTime="2025-12-09 15:46:39.483282359 +0000 UTC m=+893.802493028" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.535023 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rgwrm\" (UniqueName: \"kubernetes.io/projected/28cd4e76-f55a-4425-8bfc-47168719e9ce-kube-api-access-rgwrm\") pod \"certified-operators-jxpm6\" (UID: \"28cd4e76-f55a-4425-8bfc-47168719e9ce\") " pod="openshift-marketplace/certified-operators-jxpm6" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.535102 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28cd4e76-f55a-4425-8bfc-47168719e9ce-utilities\") pod \"certified-operators-jxpm6\" (UID: \"28cd4e76-f55a-4425-8bfc-47168719e9ce\") " pod="openshift-marketplace/certified-operators-jxpm6" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.535133 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28cd4e76-f55a-4425-8bfc-47168719e9ce-catalog-content\") pod \"certified-operators-jxpm6\" (UID: \"28cd4e76-f55a-4425-8bfc-47168719e9ce\") " pod="openshift-marketplace/certified-operators-jxpm6" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.635921 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rgwrm\" (UniqueName: \"kubernetes.io/projected/28cd4e76-f55a-4425-8bfc-47168719e9ce-kube-api-access-rgwrm\") pod \"certified-operators-jxpm6\" (UID: \"28cd4e76-f55a-4425-8bfc-47168719e9ce\") " pod="openshift-marketplace/certified-operators-jxpm6" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.635995 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28cd4e76-f55a-4425-8bfc-47168719e9ce-utilities\") pod \"certified-operators-jxpm6\" (UID: \"28cd4e76-f55a-4425-8bfc-47168719e9ce\") " pod="openshift-marketplace/certified-operators-jxpm6" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.636015 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28cd4e76-f55a-4425-8bfc-47168719e9ce-catalog-content\") pod \"certified-operators-jxpm6\" (UID: \"28cd4e76-f55a-4425-8bfc-47168719e9ce\") " pod="openshift-marketplace/certified-operators-jxpm6" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.636505 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28cd4e76-f55a-4425-8bfc-47168719e9ce-catalog-content\") pod \"certified-operators-jxpm6\" (UID: \"28cd4e76-f55a-4425-8bfc-47168719e9ce\") " pod="openshift-marketplace/certified-operators-jxpm6" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.636624 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28cd4e76-f55a-4425-8bfc-47168719e9ce-utilities\") pod \"certified-operators-jxpm6\" (UID: \"28cd4e76-f55a-4425-8bfc-47168719e9ce\") " pod="openshift-marketplace/certified-operators-jxpm6" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.677752 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rgwrm\" (UniqueName: \"kubernetes.io/projected/28cd4e76-f55a-4425-8bfc-47168719e9ce-kube-api-access-rgwrm\") pod \"certified-operators-jxpm6\" (UID: \"28cd4e76-f55a-4425-8bfc-47168719e9ce\") " pod="openshift-marketplace/certified-operators-jxpm6" Dec 09 15:46:39 crc kubenswrapper[4894]: I1209 15:46:39.734751 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jxpm6" Dec 09 15:46:40 crc kubenswrapper[4894]: I1209 15:46:40.212469 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-5ptth" event={"ID":"d85fd4c5-d08d-4bda-b8b9-aad2165e930c","Type":"ContainerStarted","Data":"02944d39184b9e0fb9f2fc61e5773729718fe682e2dacb8aba4e985b296ff617"} Dec 09 15:46:40 crc kubenswrapper[4894]: I1209 15:46:40.213178 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-5ptth" Dec 09 15:46:40 crc kubenswrapper[4894]: I1209 15:46:40.216547 4894 generic.go:334] "Generic (PLEG): container finished" podID="cd8e92cc-f08a-41f2-a439-03fe7c0bc000" containerID="05c53fdc9ba7b6efa265124d40d7e70a98f10d4df31f98b4bbf75cb12bed4f75" exitCode=0 Dec 09 15:46:40 crc kubenswrapper[4894]: I1209 15:46:40.216661 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-thsb5" event={"ID":"cd8e92cc-f08a-41f2-a439-03fe7c0bc000","Type":"ContainerDied","Data":"05c53fdc9ba7b6efa265124d40d7e70a98f10d4df31f98b4bbf75cb12bed4f75"} Dec 09 15:46:40 crc kubenswrapper[4894]: I1209 15:46:40.219461 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5kgk4" event={"ID":"4aa1a97a-d5a6-42e1-9db7-94ca1cad21ff","Type":"ContainerStarted","Data":"5d6f4d82c7afbf1bedb8428eafdf08bcbc3038737361ba031684e948fee9b8a7"} Dec 09 15:46:40 crc kubenswrapper[4894]: I1209 15:46:40.220137 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5kgk4" Dec 09 15:46:40 crc kubenswrapper[4894]: I1209 15:46:40.223688 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-nv5mv" event={"ID":"811e1bfc-20a8-4268-9bda-88dd4e1b33ad","Type":"ContainerStarted","Data":"897f1e40895f8e40c8c9cd49debe74fed7da75e6a420553c552604c2cab657ef"} Dec 09 15:46:40 crc kubenswrapper[4894]: I1209 15:46:40.223814 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-nv5mv" Dec 09 15:46:40 crc kubenswrapper[4894]: I1209 15:46:40.225400 4894 generic.go:334] "Generic (PLEG): container finished" podID="59332ca3-ed84-4e9b-b965-669f3ecbdd51" containerID="ba3f78c7c2e5fa81a4a818c1e0d985756064884f87b381464bcdfd47dcb21f52" exitCode=0 Dec 09 15:46:40 crc kubenswrapper[4894]: I1209 15:46:40.225449 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nkmzh" event={"ID":"59332ca3-ed84-4e9b-b965-669f3ecbdd51","Type":"ContainerDied","Data":"ba3f78c7c2e5fa81a4a818c1e0d985756064884f87b381464bcdfd47dcb21f52"} Dec 09 15:46:40 crc kubenswrapper[4894]: I1209 15:46:40.236278 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wpj5r" event={"ID":"faf8439c-1fc7-4316-8b58-31c5a892258b","Type":"ContainerStarted","Data":"35458b7412407f35d1d1ae5446d921a697e8e90fe6285392f369dc02e13d889a"} Dec 09 15:46:40 crc kubenswrapper[4894]: I1209 15:46:40.236936 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wpj5r" Dec 09 15:46:40 crc kubenswrapper[4894]: I1209 15:46:40.241076 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-5ptth" podStartSLOduration=7.033764562 podStartE2EDuration="39.241056651s" podCreationTimestamp="2025-12-09 15:46:01 +0000 UTC" firstStartedPulling="2025-12-09 15:46:03.202786544 +0000 UTC m=+857.521997213" lastFinishedPulling="2025-12-09 15:46:35.410078633 +0000 UTC m=+889.729289302" observedRunningTime="2025-12-09 15:46:40.236049995 +0000 UTC m=+894.555260684" watchObservedRunningTime="2025-12-09 15:46:40.241056651 +0000 UTC m=+894.560267320" Dec 09 15:46:40 crc kubenswrapper[4894]: I1209 15:46:40.243871 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-7qv58" event={"ID":"e177f09b-5191-4e61-af98-91aa6bf25fc3","Type":"ContainerStarted","Data":"6e224b8c14f274aca86c116184525c4a2b18da1f531be79f146b64fe2296b343"} Dec 09 15:46:40 crc kubenswrapper[4894]: I1209 15:46:40.246360 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-967d97867-4qzlb" Dec 09 15:46:40 crc kubenswrapper[4894]: I1209 15:46:40.249289 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-5f64f6f8bb-v2jx9" Dec 09 15:46:40 crc kubenswrapper[4894]: I1209 15:46:40.274026 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-nv5mv" podStartSLOduration=3.297503731 podStartE2EDuration="39.274005939s" podCreationTimestamp="2025-12-09 15:46:01 +0000 UTC" firstStartedPulling="2025-12-09 15:46:02.749107162 +0000 UTC m=+857.068317831" lastFinishedPulling="2025-12-09 15:46:38.72560937 +0000 UTC m=+893.044820039" observedRunningTime="2025-12-09 15:46:40.272132468 +0000 UTC m=+894.591343137" watchObservedRunningTime="2025-12-09 15:46:40.274005939 +0000 UTC m=+894.593216608" Dec 09 15:46:40 crc kubenswrapper[4894]: I1209 15:46:40.300729 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5kgk4" podStartSLOduration=2.713964417 podStartE2EDuration="39.300711369s" podCreationTimestamp="2025-12-09 15:46:01 +0000 UTC" firstStartedPulling="2025-12-09 15:46:03.063471138 +0000 UTC m=+857.382681807" lastFinishedPulling="2025-12-09 15:46:39.65021809 +0000 UTC m=+893.969428759" observedRunningTime="2025-12-09 15:46:40.294865981 +0000 UTC m=+894.614076660" watchObservedRunningTime="2025-12-09 15:46:40.300711369 +0000 UTC m=+894.619922038" Dec 09 15:46:40 crc kubenswrapper[4894]: I1209 15:46:40.319138 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-jxpm6"] Dec 09 15:46:40 crc kubenswrapper[4894]: I1209 15:46:40.388040 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wpj5r" podStartSLOduration=7.084030557 podStartE2EDuration="39.388018203s" podCreationTimestamp="2025-12-09 15:46:01 +0000 UTC" firstStartedPulling="2025-12-09 15:46:03.322314457 +0000 UTC m=+857.641525136" lastFinishedPulling="2025-12-09 15:46:35.626302113 +0000 UTC m=+889.945512782" observedRunningTime="2025-12-09 15:46:40.387592252 +0000 UTC m=+894.706802921" watchObservedRunningTime="2025-12-09 15:46:40.388018203 +0000 UTC m=+894.707228862" Dec 09 15:46:40 crc kubenswrapper[4894]: I1209 15:46:40.450845 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-7qv58" podStartSLOduration=7.356190977 podStartE2EDuration="39.450831517s" podCreationTimestamp="2025-12-09 15:46:01 +0000 UTC" firstStartedPulling="2025-12-09 15:46:03.314762674 +0000 UTC m=+857.633973353" lastFinishedPulling="2025-12-09 15:46:35.409403224 +0000 UTC m=+889.728613893" observedRunningTime="2025-12-09 15:46:40.448963166 +0000 UTC m=+894.768173845" watchObservedRunningTime="2025-12-09 15:46:40.450831517 +0000 UTC m=+894.770042186" Dec 09 15:46:41 crc kubenswrapper[4894]: I1209 15:46:41.252091 4894 generic.go:334] "Generic (PLEG): container finished" podID="28cd4e76-f55a-4425-8bfc-47168719e9ce" containerID="6151b73b4d09da97577c739e6155a69636ad8b2f13439d9165d9d3ecdc81ed10" exitCode=0 Dec 09 15:46:41 crc kubenswrapper[4894]: I1209 15:46:41.252139 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jxpm6" event={"ID":"28cd4e76-f55a-4425-8bfc-47168719e9ce","Type":"ContainerDied","Data":"6151b73b4d09da97577c739e6155a69636ad8b2f13439d9165d9d3ecdc81ed10"} Dec 09 15:46:41 crc kubenswrapper[4894]: I1209 15:46:41.252619 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jxpm6" event={"ID":"28cd4e76-f55a-4425-8bfc-47168719e9ce","Type":"ContainerStarted","Data":"995add32cfe915bd9f4d261270ee2c83bcb627031c07565389aa06083101ddde"} Dec 09 15:46:41 crc kubenswrapper[4894]: I1209 15:46:41.256620 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-thsb5" event={"ID":"cd8e92cc-f08a-41f2-a439-03fe7c0bc000","Type":"ContainerStarted","Data":"d6d3fd9da932c133411520e7883820893368eb45c9dbc8282037522cb26125c5"} Dec 09 15:46:41 crc kubenswrapper[4894]: I1209 15:46:41.258015 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-7qv58" Dec 09 15:46:41 crc kubenswrapper[4894]: I1209 15:46:41.287354 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-thsb5" podStartSLOduration=33.435438482 podStartE2EDuration="37.287333731s" podCreationTimestamp="2025-12-09 15:46:04 +0000 UTC" firstStartedPulling="2025-12-09 15:46:36.788115869 +0000 UTC m=+891.107326538" lastFinishedPulling="2025-12-09 15:46:40.640011118 +0000 UTC m=+894.959221787" observedRunningTime="2025-12-09 15:46:41.286283523 +0000 UTC m=+895.605494202" watchObservedRunningTime="2025-12-09 15:46:41.287333731 +0000 UTC m=+895.606544400" Dec 09 15:46:42 crc kubenswrapper[4894]: I1209 15:46:42.019855 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-5b5fd79c9c-5ptth" Dec 09 15:46:42 crc kubenswrapper[4894]: I1209 15:46:42.157395 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-9d58d64bc-wxcz2" Dec 09 15:46:42 crc kubenswrapper[4894]: I1209 15:46:42.177500 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-697bc559fc-7qv58" Dec 09 15:46:42 crc kubenswrapper[4894]: I1209 15:46:42.200231 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-78f8948974-wpj5r" Dec 09 15:46:42 crc kubenswrapper[4894]: I1209 15:46:42.257568 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:46:42 crc kubenswrapper[4894]: I1209 15:46:42.257625 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:46:42 crc kubenswrapper[4894]: I1209 15:46:42.529353 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-58d5ff84df-frzr9" Dec 09 15:46:42 crc kubenswrapper[4894]: I1209 15:46:42.571498 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8" Dec 09 15:46:42 crc kubenswrapper[4894]: I1209 15:46:42.592902 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-667bd8d554-c9v7t" Dec 09 15:46:44 crc kubenswrapper[4894]: I1209 15:46:44.427130 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-668858c49-xgc5f" Dec 09 15:46:45 crc kubenswrapper[4894]: I1209 15:46:45.240675 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-thsb5" Dec 09 15:46:45 crc kubenswrapper[4894]: I1209 15:46:45.240735 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-thsb5" Dec 09 15:46:45 crc kubenswrapper[4894]: I1209 15:46:45.297340 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nkmzh" event={"ID":"59332ca3-ed84-4e9b-b965-669f3ecbdd51","Type":"ContainerStarted","Data":"40214f4e04ce8a3a6e05391cf9648704072e42d0861fa598738f17eaf74b3906"} Dec 09 15:46:45 crc kubenswrapper[4894]: I1209 15:46:45.298951 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6fdzd" event={"ID":"7700fe8f-1bda-4f92-81ff-04463e0e4fdf","Type":"ContainerStarted","Data":"0d0414c990908f97e0e9b4a664c7748160c65280f3ef13f39c9fbf5525408c1c"} Dec 09 15:46:45 crc kubenswrapper[4894]: I1209 15:46:45.300339 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" event={"ID":"9557c24e-8469-46ec-9bab-04e337606beb","Type":"ContainerStarted","Data":"44dc8ee65ac5ca6812a195f34da6e15eee4edfc889dff4f3e6df4de7875f0837"} Dec 09 15:46:45 crc kubenswrapper[4894]: I1209 15:46:45.300379 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" event={"ID":"9557c24e-8469-46ec-9bab-04e337606beb","Type":"ContainerStarted","Data":"ecc87a976da66492334e8cc3dc15c2179ccbdeb6f40970a7fa23ebde2f3541e3"} Dec 09 15:46:45 crc kubenswrapper[4894]: I1209 15:46:45.300831 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" Dec 09 15:46:45 crc kubenswrapper[4894]: I1209 15:46:45.326006 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-nkmzh" podStartSLOduration=29.966719168 podStartE2EDuration="37.325984206s" podCreationTimestamp="2025-12-09 15:46:08 +0000 UTC" firstStartedPulling="2025-12-09 15:46:36.788238662 +0000 UTC m=+891.107449331" lastFinishedPulling="2025-12-09 15:46:44.1475037 +0000 UTC m=+898.466714369" observedRunningTime="2025-12-09 15:46:45.315713068 +0000 UTC m=+899.634923727" watchObservedRunningTime="2025-12-09 15:46:45.325984206 +0000 UTC m=+899.645194875" Dec 09 15:46:45 crc kubenswrapper[4894]: I1209 15:46:45.337754 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-668c99d594-6fdzd" podStartSLOduration=2.44062944 podStartE2EDuration="43.337738072s" podCreationTimestamp="2025-12-09 15:46:02 +0000 UTC" firstStartedPulling="2025-12-09 15:46:03.314484196 +0000 UTC m=+857.633694865" lastFinishedPulling="2025-12-09 15:46:44.211592828 +0000 UTC m=+898.530803497" observedRunningTime="2025-12-09 15:46:45.334700431 +0000 UTC m=+899.653911100" watchObservedRunningTime="2025-12-09 15:46:45.337738072 +0000 UTC m=+899.656948741" Dec 09 15:46:45 crc kubenswrapper[4894]: I1209 15:46:45.370665 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" podStartSLOduration=37.078705738 podStartE2EDuration="44.3706315s" podCreationTimestamp="2025-12-09 15:46:01 +0000 UTC" firstStartedPulling="2025-12-09 15:46:36.85566505 +0000 UTC m=+891.174875719" lastFinishedPulling="2025-12-09 15:46:44.147590812 +0000 UTC m=+898.466801481" observedRunningTime="2025-12-09 15:46:45.36396117 +0000 UTC m=+899.683171849" watchObservedRunningTime="2025-12-09 15:46:45.3706315 +0000 UTC m=+899.689842169" Dec 09 15:46:46 crc kubenswrapper[4894]: I1209 15:46:46.293203 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-thsb5" podUID="cd8e92cc-f08a-41f2-a439-03fe7c0bc000" containerName="registry-server" probeResult="failure" output=< Dec 09 15:46:46 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 09 15:46:46 crc kubenswrapper[4894]: > Dec 09 15:46:47 crc kubenswrapper[4894]: I1209 15:46:47.316010 4894 generic.go:334] "Generic (PLEG): container finished" podID="28cd4e76-f55a-4425-8bfc-47168719e9ce" containerID="e11e7f8c4355eed41893047f3b20d4340afd875f1211935434ca1afdc0c3f1e9" exitCode=0 Dec 09 15:46:47 crc kubenswrapper[4894]: I1209 15:46:47.316060 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jxpm6" event={"ID":"28cd4e76-f55a-4425-8bfc-47168719e9ce","Type":"ContainerDied","Data":"e11e7f8c4355eed41893047f3b20d4340afd875f1211935434ca1afdc0c3f1e9"} Dec 09 15:46:47 crc kubenswrapper[4894]: I1209 15:46:47.633923 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-78d48bff9d-mxngk" Dec 09 15:46:48 crc kubenswrapper[4894]: I1209 15:46:48.325755 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jxpm6" event={"ID":"28cd4e76-f55a-4425-8bfc-47168719e9ce","Type":"ContainerStarted","Data":"2050649d3ca13977c91a47c93bdf2250dc3f3f08f2dd67bd1e7bc87ff7879559"} Dec 09 15:46:48 crc kubenswrapper[4894]: I1209 15:46:48.346286 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-jxpm6" podStartSLOduration=4.947902118 podStartE2EDuration="9.346258651s" podCreationTimestamp="2025-12-09 15:46:39 +0000 UTC" firstStartedPulling="2025-12-09 15:46:43.596891884 +0000 UTC m=+897.916102553" lastFinishedPulling="2025-12-09 15:46:47.995248417 +0000 UTC m=+902.314459086" observedRunningTime="2025-12-09 15:46:48.344532525 +0000 UTC m=+902.663743194" watchObservedRunningTime="2025-12-09 15:46:48.346258651 +0000 UTC m=+902.665469320" Dec 09 15:46:48 crc kubenswrapper[4894]: I1209 15:46:48.454147 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-nkmzh" Dec 09 15:46:48 crc kubenswrapper[4894]: I1209 15:46:48.454224 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-nkmzh" Dec 09 15:46:48 crc kubenswrapper[4894]: I1209 15:46:48.506455 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-nkmzh" Dec 09 15:46:49 crc kubenswrapper[4894]: I1209 15:46:49.392034 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-nkmzh" Dec 09 15:46:49 crc kubenswrapper[4894]: I1209 15:46:49.735115 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-jxpm6" Dec 09 15:46:49 crc kubenswrapper[4894]: I1209 15:46:49.735602 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-jxpm6" Dec 09 15:46:50 crc kubenswrapper[4894]: I1209 15:46:50.572361 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nkmzh"] Dec 09 15:46:50 crc kubenswrapper[4894]: I1209 15:46:50.790009 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/certified-operators-jxpm6" podUID="28cd4e76-f55a-4425-8bfc-47168719e9ce" containerName="registry-server" probeResult="failure" output=< Dec 09 15:46:50 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 09 15:46:50 crc kubenswrapper[4894]: > Dec 09 15:46:51 crc kubenswrapper[4894]: I1209 15:46:51.347477 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-nkmzh" podUID="59332ca3-ed84-4e9b-b965-669f3ecbdd51" containerName="registry-server" containerID="cri-o://40214f4e04ce8a3a6e05391cf9648704072e42d0861fa598738f17eaf74b3906" gracePeriod=2 Dec 09 15:46:51 crc kubenswrapper[4894]: I1209 15:46:51.526873 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-5697bb5779-nv5mv" Dec 09 15:46:51 crc kubenswrapper[4894]: I1209 15:46:51.987333 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-7765d96ddf-5kgk4" Dec 09 15:46:52 crc kubenswrapper[4894]: I1209 15:46:52.977839 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-pb2wl"] Dec 09 15:46:52 crc kubenswrapper[4894]: I1209 15:46:52.979656 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pb2wl" Dec 09 15:46:52 crc kubenswrapper[4894]: I1209 15:46:52.999706 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pb2wl"] Dec 09 15:46:53 crc kubenswrapper[4894]: I1209 15:46:53.048235 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cc7356e-a563-4343-9d84-7650b75b3f34-utilities\") pod \"redhat-marketplace-pb2wl\" (UID: \"4cc7356e-a563-4343-9d84-7650b75b3f34\") " pod="openshift-marketplace/redhat-marketplace-pb2wl" Dec 09 15:46:53 crc kubenswrapper[4894]: I1209 15:46:53.048326 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cc7356e-a563-4343-9d84-7650b75b3f34-catalog-content\") pod \"redhat-marketplace-pb2wl\" (UID: \"4cc7356e-a563-4343-9d84-7650b75b3f34\") " pod="openshift-marketplace/redhat-marketplace-pb2wl" Dec 09 15:46:53 crc kubenswrapper[4894]: I1209 15:46:53.048387 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5c5sl\" (UniqueName: \"kubernetes.io/projected/4cc7356e-a563-4343-9d84-7650b75b3f34-kube-api-access-5c5sl\") pod \"redhat-marketplace-pb2wl\" (UID: \"4cc7356e-a563-4343-9d84-7650b75b3f34\") " pod="openshift-marketplace/redhat-marketplace-pb2wl" Dec 09 15:46:53 crc kubenswrapper[4894]: I1209 15:46:53.150084 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cc7356e-a563-4343-9d84-7650b75b3f34-catalog-content\") pod \"redhat-marketplace-pb2wl\" (UID: \"4cc7356e-a563-4343-9d84-7650b75b3f34\") " pod="openshift-marketplace/redhat-marketplace-pb2wl" Dec 09 15:46:53 crc kubenswrapper[4894]: I1209 15:46:53.150146 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5c5sl\" (UniqueName: \"kubernetes.io/projected/4cc7356e-a563-4343-9d84-7650b75b3f34-kube-api-access-5c5sl\") pod \"redhat-marketplace-pb2wl\" (UID: \"4cc7356e-a563-4343-9d84-7650b75b3f34\") " pod="openshift-marketplace/redhat-marketplace-pb2wl" Dec 09 15:46:53 crc kubenswrapper[4894]: I1209 15:46:53.150193 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cc7356e-a563-4343-9d84-7650b75b3f34-utilities\") pod \"redhat-marketplace-pb2wl\" (UID: \"4cc7356e-a563-4343-9d84-7650b75b3f34\") " pod="openshift-marketplace/redhat-marketplace-pb2wl" Dec 09 15:46:53 crc kubenswrapper[4894]: I1209 15:46:53.150806 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cc7356e-a563-4343-9d84-7650b75b3f34-catalog-content\") pod \"redhat-marketplace-pb2wl\" (UID: \"4cc7356e-a563-4343-9d84-7650b75b3f34\") " pod="openshift-marketplace/redhat-marketplace-pb2wl" Dec 09 15:46:53 crc kubenswrapper[4894]: I1209 15:46:53.150812 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cc7356e-a563-4343-9d84-7650b75b3f34-utilities\") pod \"redhat-marketplace-pb2wl\" (UID: \"4cc7356e-a563-4343-9d84-7650b75b3f34\") " pod="openshift-marketplace/redhat-marketplace-pb2wl" Dec 09 15:46:53 crc kubenswrapper[4894]: I1209 15:46:53.167908 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5c5sl\" (UniqueName: \"kubernetes.io/projected/4cc7356e-a563-4343-9d84-7650b75b3f34-kube-api-access-5c5sl\") pod \"redhat-marketplace-pb2wl\" (UID: \"4cc7356e-a563-4343-9d84-7650b75b3f34\") " pod="openshift-marketplace/redhat-marketplace-pb2wl" Dec 09 15:46:53 crc kubenswrapper[4894]: I1209 15:46:53.359039 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pb2wl" Dec 09 15:46:53 crc kubenswrapper[4894]: I1209 15:46:53.598606 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-pb2wl"] Dec 09 15:46:53 crc kubenswrapper[4894]: I1209 15:46:53.908469 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-694d6cfbd667tpw" Dec 09 15:46:54 crc kubenswrapper[4894]: I1209 15:46:54.370733 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pb2wl" event={"ID":"4cc7356e-a563-4343-9d84-7650b75b3f34","Type":"ContainerStarted","Data":"df0bb7e8fc52aad5edc742dfdc431bd6ada55dcda8c85ebc3ada004e83a40685"} Dec 09 15:46:55 crc kubenswrapper[4894]: I1209 15:46:55.283841 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-thsb5" Dec 09 15:46:55 crc kubenswrapper[4894]: I1209 15:46:55.325397 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-thsb5" Dec 09 15:46:55 crc kubenswrapper[4894]: I1209 15:46:55.386835 4894 generic.go:334] "Generic (PLEG): container finished" podID="59332ca3-ed84-4e9b-b965-669f3ecbdd51" containerID="40214f4e04ce8a3a6e05391cf9648704072e42d0861fa598738f17eaf74b3906" exitCode=0 Dec 09 15:46:55 crc kubenswrapper[4894]: I1209 15:46:55.386922 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nkmzh" event={"ID":"59332ca3-ed84-4e9b-b965-669f3ecbdd51","Type":"ContainerDied","Data":"40214f4e04ce8a3a6e05391cf9648704072e42d0861fa598738f17eaf74b3906"} Dec 09 15:46:56 crc kubenswrapper[4894]: I1209 15:46:56.189552 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nkmzh" Dec 09 15:46:56 crc kubenswrapper[4894]: I1209 15:46:56.291589 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59332ca3-ed84-4e9b-b965-669f3ecbdd51-catalog-content\") pod \"59332ca3-ed84-4e9b-b965-669f3ecbdd51\" (UID: \"59332ca3-ed84-4e9b-b965-669f3ecbdd51\") " Dec 09 15:46:56 crc kubenswrapper[4894]: I1209 15:46:56.291726 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59332ca3-ed84-4e9b-b965-669f3ecbdd51-utilities\") pod \"59332ca3-ed84-4e9b-b965-669f3ecbdd51\" (UID: \"59332ca3-ed84-4e9b-b965-669f3ecbdd51\") " Dec 09 15:46:56 crc kubenswrapper[4894]: I1209 15:46:56.291785 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wbjn7\" (UniqueName: \"kubernetes.io/projected/59332ca3-ed84-4e9b-b965-669f3ecbdd51-kube-api-access-wbjn7\") pod \"59332ca3-ed84-4e9b-b965-669f3ecbdd51\" (UID: \"59332ca3-ed84-4e9b-b965-669f3ecbdd51\") " Dec 09 15:46:56 crc kubenswrapper[4894]: I1209 15:46:56.299961 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59332ca3-ed84-4e9b-b965-669f3ecbdd51-utilities" (OuterVolumeSpecName: "utilities") pod "59332ca3-ed84-4e9b-b965-669f3ecbdd51" (UID: "59332ca3-ed84-4e9b-b965-669f3ecbdd51"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:46:56 crc kubenswrapper[4894]: I1209 15:46:56.302879 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59332ca3-ed84-4e9b-b965-669f3ecbdd51-kube-api-access-wbjn7" (OuterVolumeSpecName: "kube-api-access-wbjn7") pod "59332ca3-ed84-4e9b-b965-669f3ecbdd51" (UID: "59332ca3-ed84-4e9b-b965-669f3ecbdd51"). InnerVolumeSpecName "kube-api-access-wbjn7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:46:56 crc kubenswrapper[4894]: I1209 15:46:56.343492 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59332ca3-ed84-4e9b-b965-669f3ecbdd51-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59332ca3-ed84-4e9b-b965-669f3ecbdd51" (UID: "59332ca3-ed84-4e9b-b965-669f3ecbdd51"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:46:56 crc kubenswrapper[4894]: I1209 15:46:56.393673 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59332ca3-ed84-4e9b-b965-669f3ecbdd51-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 15:46:56 crc kubenswrapper[4894]: I1209 15:46:56.393718 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59332ca3-ed84-4e9b-b965-669f3ecbdd51-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 15:46:56 crc kubenswrapper[4894]: I1209 15:46:56.393731 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wbjn7\" (UniqueName: \"kubernetes.io/projected/59332ca3-ed84-4e9b-b965-669f3ecbdd51-kube-api-access-wbjn7\") on node \"crc\" DevicePath \"\"" Dec 09 15:46:56 crc kubenswrapper[4894]: I1209 15:46:56.398237 4894 generic.go:334] "Generic (PLEG): container finished" podID="4cc7356e-a563-4343-9d84-7650b75b3f34" containerID="d367cf4408e633f7ad5215745e14cf7c63996a9f4fdafdfdecbd17b904878809" exitCode=0 Dec 09 15:46:56 crc kubenswrapper[4894]: I1209 15:46:56.398342 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pb2wl" event={"ID":"4cc7356e-a563-4343-9d84-7650b75b3f34","Type":"ContainerDied","Data":"d367cf4408e633f7ad5215745e14cf7c63996a9f4fdafdfdecbd17b904878809"} Dec 09 15:46:56 crc kubenswrapper[4894]: I1209 15:46:56.402487 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-nkmzh" event={"ID":"59332ca3-ed84-4e9b-b965-669f3ecbdd51","Type":"ContainerDied","Data":"3b9b10724e1e536853a73416337b75d496ac25049377f78927aaeae67da39e14"} Dec 09 15:46:56 crc kubenswrapper[4894]: I1209 15:46:56.402534 4894 scope.go:117] "RemoveContainer" containerID="40214f4e04ce8a3a6e05391cf9648704072e42d0861fa598738f17eaf74b3906" Dec 09 15:46:56 crc kubenswrapper[4894]: I1209 15:46:56.402654 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-nkmzh" Dec 09 15:46:56 crc kubenswrapper[4894]: I1209 15:46:56.435405 4894 scope.go:117] "RemoveContainer" containerID="ba3f78c7c2e5fa81a4a818c1e0d985756064884f87b381464bcdfd47dcb21f52" Dec 09 15:46:56 crc kubenswrapper[4894]: I1209 15:46:56.460149 4894 scope.go:117] "RemoveContainer" containerID="61bda99fa3c7b7c2fd7a05cdec3621c434d888aee71787e243e1cf4956e64dfc" Dec 09 15:46:56 crc kubenswrapper[4894]: I1209 15:46:56.478026 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-nkmzh"] Dec 09 15:46:56 crc kubenswrapper[4894]: I1209 15:46:56.502721 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-nkmzh"] Dec 09 15:46:57 crc kubenswrapper[4894]: I1209 15:46:57.565558 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-thsb5"] Dec 09 15:46:57 crc kubenswrapper[4894]: I1209 15:46:57.566037 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-thsb5" podUID="cd8e92cc-f08a-41f2-a439-03fe7c0bc000" containerName="registry-server" containerID="cri-o://d6d3fd9da932c133411520e7883820893368eb45c9dbc8282037522cb26125c5" gracePeriod=2 Dec 09 15:46:58 crc kubenswrapper[4894]: I1209 15:46:58.115554 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59332ca3-ed84-4e9b-b965-669f3ecbdd51" path="/var/lib/kubelet/pods/59332ca3-ed84-4e9b-b965-669f3ecbdd51/volumes" Dec 09 15:46:58 crc kubenswrapper[4894]: I1209 15:46:58.420049 4894 generic.go:334] "Generic (PLEG): container finished" podID="cd8e92cc-f08a-41f2-a439-03fe7c0bc000" containerID="d6d3fd9da932c133411520e7883820893368eb45c9dbc8282037522cb26125c5" exitCode=0 Dec 09 15:46:58 crc kubenswrapper[4894]: I1209 15:46:58.420096 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-thsb5" event={"ID":"cd8e92cc-f08a-41f2-a439-03fe7c0bc000","Type":"ContainerDied","Data":"d6d3fd9da932c133411520e7883820893368eb45c9dbc8282037522cb26125c5"} Dec 09 15:46:59 crc kubenswrapper[4894]: I1209 15:46:59.791095 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-jxpm6" Dec 09 15:46:59 crc kubenswrapper[4894]: I1209 15:46:59.847685 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-jxpm6" Dec 09 15:47:01 crc kubenswrapper[4894]: I1209 15:47:01.565000 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jxpm6"] Dec 09 15:47:01 crc kubenswrapper[4894]: I1209 15:47:01.565737 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-jxpm6" podUID="28cd4e76-f55a-4425-8bfc-47168719e9ce" containerName="registry-server" containerID="cri-o://2050649d3ca13977c91a47c93bdf2250dc3f3f08f2dd67bd1e7bc87ff7879559" gracePeriod=2 Dec 09 15:47:02 crc kubenswrapper[4894]: I1209 15:47:02.402675 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-thsb5" Dec 09 15:47:02 crc kubenswrapper[4894]: I1209 15:47:02.460333 4894 generic.go:334] "Generic (PLEG): container finished" podID="28cd4e76-f55a-4425-8bfc-47168719e9ce" containerID="2050649d3ca13977c91a47c93bdf2250dc3f3f08f2dd67bd1e7bc87ff7879559" exitCode=0 Dec 09 15:47:02 crc kubenswrapper[4894]: I1209 15:47:02.460405 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jxpm6" event={"ID":"28cd4e76-f55a-4425-8bfc-47168719e9ce","Type":"ContainerDied","Data":"2050649d3ca13977c91a47c93bdf2250dc3f3f08f2dd67bd1e7bc87ff7879559"} Dec 09 15:47:02 crc kubenswrapper[4894]: I1209 15:47:02.462558 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-thsb5" event={"ID":"cd8e92cc-f08a-41f2-a439-03fe7c0bc000","Type":"ContainerDied","Data":"d19e1a38a6a18767e020576bb99489f7498a8feaf99d252d6a03be3715eff920"} Dec 09 15:47:02 crc kubenswrapper[4894]: I1209 15:47:02.462708 4894 scope.go:117] "RemoveContainer" containerID="d6d3fd9da932c133411520e7883820893368eb45c9dbc8282037522cb26125c5" Dec 09 15:47:02 crc kubenswrapper[4894]: I1209 15:47:02.462789 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-thsb5" Dec 09 15:47:02 crc kubenswrapper[4894]: I1209 15:47:02.484711 4894 scope.go:117] "RemoveContainer" containerID="05c53fdc9ba7b6efa265124d40d7e70a98f10d4df31f98b4bbf75cb12bed4f75" Dec 09 15:47:02 crc kubenswrapper[4894]: I1209 15:47:02.485319 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd8e92cc-f08a-41f2-a439-03fe7c0bc000-utilities\") pod \"cd8e92cc-f08a-41f2-a439-03fe7c0bc000\" (UID: \"cd8e92cc-f08a-41f2-a439-03fe7c0bc000\") " Dec 09 15:47:02 crc kubenswrapper[4894]: I1209 15:47:02.485374 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfbqn\" (UniqueName: \"kubernetes.io/projected/cd8e92cc-f08a-41f2-a439-03fe7c0bc000-kube-api-access-tfbqn\") pod \"cd8e92cc-f08a-41f2-a439-03fe7c0bc000\" (UID: \"cd8e92cc-f08a-41f2-a439-03fe7c0bc000\") " Dec 09 15:47:02 crc kubenswrapper[4894]: I1209 15:47:02.485401 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd8e92cc-f08a-41f2-a439-03fe7c0bc000-catalog-content\") pod \"cd8e92cc-f08a-41f2-a439-03fe7c0bc000\" (UID: \"cd8e92cc-f08a-41f2-a439-03fe7c0bc000\") " Dec 09 15:47:02 crc kubenswrapper[4894]: I1209 15:47:02.486468 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd8e92cc-f08a-41f2-a439-03fe7c0bc000-utilities" (OuterVolumeSpecName: "utilities") pod "cd8e92cc-f08a-41f2-a439-03fe7c0bc000" (UID: "cd8e92cc-f08a-41f2-a439-03fe7c0bc000"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:47:02 crc kubenswrapper[4894]: I1209 15:47:02.492949 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd8e92cc-f08a-41f2-a439-03fe7c0bc000-kube-api-access-tfbqn" (OuterVolumeSpecName: "kube-api-access-tfbqn") pod "cd8e92cc-f08a-41f2-a439-03fe7c0bc000" (UID: "cd8e92cc-f08a-41f2-a439-03fe7c0bc000"). InnerVolumeSpecName "kube-api-access-tfbqn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:47:02 crc kubenswrapper[4894]: I1209 15:47:02.502272 4894 scope.go:117] "RemoveContainer" containerID="3d74115cfd0e8ae4a5fb19fa386ac82eccb6627741cbd6107bcfd721cd6198e4" Dec 09 15:47:02 crc kubenswrapper[4894]: I1209 15:47:02.586543 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfbqn\" (UniqueName: \"kubernetes.io/projected/cd8e92cc-f08a-41f2-a439-03fe7c0bc000-kube-api-access-tfbqn\") on node \"crc\" DevicePath \"\"" Dec 09 15:47:02 crc kubenswrapper[4894]: I1209 15:47:02.586574 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd8e92cc-f08a-41f2-a439-03fe7c0bc000-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 15:47:02 crc kubenswrapper[4894]: I1209 15:47:02.598361 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cd8e92cc-f08a-41f2-a439-03fe7c0bc000-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cd8e92cc-f08a-41f2-a439-03fe7c0bc000" (UID: "cd8e92cc-f08a-41f2-a439-03fe7c0bc000"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:47:02 crc kubenswrapper[4894]: I1209 15:47:02.687323 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd8e92cc-f08a-41f2-a439-03fe7c0bc000-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 15:47:02 crc kubenswrapper[4894]: I1209 15:47:02.806990 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-thsb5"] Dec 09 15:47:02 crc kubenswrapper[4894]: I1209 15:47:02.822122 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-thsb5"] Dec 09 15:47:03 crc kubenswrapper[4894]: I1209 15:47:03.016014 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jxpm6" Dec 09 15:47:03 crc kubenswrapper[4894]: I1209 15:47:03.091321 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rgwrm\" (UniqueName: \"kubernetes.io/projected/28cd4e76-f55a-4425-8bfc-47168719e9ce-kube-api-access-rgwrm\") pod \"28cd4e76-f55a-4425-8bfc-47168719e9ce\" (UID: \"28cd4e76-f55a-4425-8bfc-47168719e9ce\") " Dec 09 15:47:03 crc kubenswrapper[4894]: I1209 15:47:03.091446 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28cd4e76-f55a-4425-8bfc-47168719e9ce-utilities\") pod \"28cd4e76-f55a-4425-8bfc-47168719e9ce\" (UID: \"28cd4e76-f55a-4425-8bfc-47168719e9ce\") " Dec 09 15:47:03 crc kubenswrapper[4894]: I1209 15:47:03.091518 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28cd4e76-f55a-4425-8bfc-47168719e9ce-catalog-content\") pod \"28cd4e76-f55a-4425-8bfc-47168719e9ce\" (UID: \"28cd4e76-f55a-4425-8bfc-47168719e9ce\") " Dec 09 15:47:03 crc kubenswrapper[4894]: I1209 15:47:03.092207 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28cd4e76-f55a-4425-8bfc-47168719e9ce-utilities" (OuterVolumeSpecName: "utilities") pod "28cd4e76-f55a-4425-8bfc-47168719e9ce" (UID: "28cd4e76-f55a-4425-8bfc-47168719e9ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:47:03 crc kubenswrapper[4894]: I1209 15:47:03.095309 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28cd4e76-f55a-4425-8bfc-47168719e9ce-kube-api-access-rgwrm" (OuterVolumeSpecName: "kube-api-access-rgwrm") pod "28cd4e76-f55a-4425-8bfc-47168719e9ce" (UID: "28cd4e76-f55a-4425-8bfc-47168719e9ce"). InnerVolumeSpecName "kube-api-access-rgwrm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:47:03 crc kubenswrapper[4894]: I1209 15:47:03.147312 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/28cd4e76-f55a-4425-8bfc-47168719e9ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "28cd4e76-f55a-4425-8bfc-47168719e9ce" (UID: "28cd4e76-f55a-4425-8bfc-47168719e9ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:47:03 crc kubenswrapper[4894]: I1209 15:47:03.193081 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rgwrm\" (UniqueName: \"kubernetes.io/projected/28cd4e76-f55a-4425-8bfc-47168719e9ce-kube-api-access-rgwrm\") on node \"crc\" DevicePath \"\"" Dec 09 15:47:03 crc kubenswrapper[4894]: I1209 15:47:03.193126 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/28cd4e76-f55a-4425-8bfc-47168719e9ce-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 15:47:03 crc kubenswrapper[4894]: I1209 15:47:03.193140 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/28cd4e76-f55a-4425-8bfc-47168719e9ce-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 15:47:03 crc kubenswrapper[4894]: I1209 15:47:03.471237 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-jxpm6" event={"ID":"28cd4e76-f55a-4425-8bfc-47168719e9ce","Type":"ContainerDied","Data":"995add32cfe915bd9f4d261270ee2c83bcb627031c07565389aa06083101ddde"} Dec 09 15:47:03 crc kubenswrapper[4894]: I1209 15:47:03.471277 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-jxpm6" Dec 09 15:47:03 crc kubenswrapper[4894]: I1209 15:47:03.471293 4894 scope.go:117] "RemoveContainer" containerID="2050649d3ca13977c91a47c93bdf2250dc3f3f08f2dd67bd1e7bc87ff7879559" Dec 09 15:47:03 crc kubenswrapper[4894]: I1209 15:47:03.475265 4894 generic.go:334] "Generic (PLEG): container finished" podID="4cc7356e-a563-4343-9d84-7650b75b3f34" containerID="6410ddca9172568afb45375e59361b3df8dba0ccc28566f3b35a14f06ca50fcf" exitCode=0 Dec 09 15:47:03 crc kubenswrapper[4894]: I1209 15:47:03.475313 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pb2wl" event={"ID":"4cc7356e-a563-4343-9d84-7650b75b3f34","Type":"ContainerDied","Data":"6410ddca9172568afb45375e59361b3df8dba0ccc28566f3b35a14f06ca50fcf"} Dec 09 15:47:03 crc kubenswrapper[4894]: I1209 15:47:03.489948 4894 scope.go:117] "RemoveContainer" containerID="e11e7f8c4355eed41893047f3b20d4340afd875f1211935434ca1afdc0c3f1e9" Dec 09 15:47:03 crc kubenswrapper[4894]: I1209 15:47:03.519406 4894 scope.go:117] "RemoveContainer" containerID="6151b73b4d09da97577c739e6155a69636ad8b2f13439d9165d9d3ecdc81ed10" Dec 09 15:47:03 crc kubenswrapper[4894]: I1209 15:47:03.520203 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-jxpm6"] Dec 09 15:47:03 crc kubenswrapper[4894]: I1209 15:47:03.529265 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-jxpm6"] Dec 09 15:47:04 crc kubenswrapper[4894]: I1209 15:47:04.115443 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28cd4e76-f55a-4425-8bfc-47168719e9ce" path="/var/lib/kubelet/pods/28cd4e76-f55a-4425-8bfc-47168719e9ce/volumes" Dec 09 15:47:04 crc kubenswrapper[4894]: I1209 15:47:04.117671 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd8e92cc-f08a-41f2-a439-03fe7c0bc000" path="/var/lib/kubelet/pods/cd8e92cc-f08a-41f2-a439-03fe7c0bc000/volumes" Dec 09 15:47:04 crc kubenswrapper[4894]: I1209 15:47:04.488094 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pb2wl" event={"ID":"4cc7356e-a563-4343-9d84-7650b75b3f34","Type":"ContainerStarted","Data":"86991a71d70961601eb92e98464d89f445932bd35ad8d61346ff3df108660949"} Dec 09 15:47:04 crc kubenswrapper[4894]: I1209 15:47:04.510165 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-pb2wl" podStartSLOduration=4.818950212 podStartE2EDuration="12.51015044s" podCreationTimestamp="2025-12-09 15:46:52 +0000 UTC" firstStartedPulling="2025-12-09 15:46:56.400625163 +0000 UTC m=+910.719835842" lastFinishedPulling="2025-12-09 15:47:04.091825391 +0000 UTC m=+918.411036070" observedRunningTime="2025-12-09 15:47:04.506837171 +0000 UTC m=+918.826047850" watchObservedRunningTime="2025-12-09 15:47:04.51015044 +0000 UTC m=+918.829361109" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.678050 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-4n4xr"] Dec 09 15:47:10 crc kubenswrapper[4894]: E1209 15:47:10.678823 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28cd4e76-f55a-4425-8bfc-47168719e9ce" containerName="extract-content" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.678838 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="28cd4e76-f55a-4425-8bfc-47168719e9ce" containerName="extract-content" Dec 09 15:47:10 crc kubenswrapper[4894]: E1209 15:47:10.678866 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59332ca3-ed84-4e9b-b965-669f3ecbdd51" containerName="registry-server" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.678873 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="59332ca3-ed84-4e9b-b965-669f3ecbdd51" containerName="registry-server" Dec 09 15:47:10 crc kubenswrapper[4894]: E1209 15:47:10.678902 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28cd4e76-f55a-4425-8bfc-47168719e9ce" containerName="registry-server" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.678910 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="28cd4e76-f55a-4425-8bfc-47168719e9ce" containerName="registry-server" Dec 09 15:47:10 crc kubenswrapper[4894]: E1209 15:47:10.678926 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd8e92cc-f08a-41f2-a439-03fe7c0bc000" containerName="registry-server" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.678934 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd8e92cc-f08a-41f2-a439-03fe7c0bc000" containerName="registry-server" Dec 09 15:47:10 crc kubenswrapper[4894]: E1209 15:47:10.678946 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd8e92cc-f08a-41f2-a439-03fe7c0bc000" containerName="extract-utilities" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.678953 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd8e92cc-f08a-41f2-a439-03fe7c0bc000" containerName="extract-utilities" Dec 09 15:47:10 crc kubenswrapper[4894]: E1209 15:47:10.678967 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59332ca3-ed84-4e9b-b965-669f3ecbdd51" containerName="extract-utilities" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.678975 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="59332ca3-ed84-4e9b-b965-669f3ecbdd51" containerName="extract-utilities" Dec 09 15:47:10 crc kubenswrapper[4894]: E1209 15:47:10.678984 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28cd4e76-f55a-4425-8bfc-47168719e9ce" containerName="extract-utilities" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.678990 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="28cd4e76-f55a-4425-8bfc-47168719e9ce" containerName="extract-utilities" Dec 09 15:47:10 crc kubenswrapper[4894]: E1209 15:47:10.679000 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59332ca3-ed84-4e9b-b965-669f3ecbdd51" containerName="extract-content" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.679007 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="59332ca3-ed84-4e9b-b965-669f3ecbdd51" containerName="extract-content" Dec 09 15:47:10 crc kubenswrapper[4894]: E1209 15:47:10.679021 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd8e92cc-f08a-41f2-a439-03fe7c0bc000" containerName="extract-content" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.679027 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd8e92cc-f08a-41f2-a439-03fe7c0bc000" containerName="extract-content" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.679190 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="59332ca3-ed84-4e9b-b965-669f3ecbdd51" containerName="registry-server" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.679210 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="28cd4e76-f55a-4425-8bfc-47168719e9ce" containerName="registry-server" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.679237 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd8e92cc-f08a-41f2-a439-03fe7c0bc000" containerName="registry-server" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.680261 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bb9d8bd9-4n4xr" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.682562 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.682851 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.682972 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.683148 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-ps9h2" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.704030 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-4n4xr"] Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.742132 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-cjt4j"] Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.750498 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f854695bc-cjt4j" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.756118 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.771063 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-cjt4j"] Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.818566 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqgxz\" (UniqueName: \"kubernetes.io/projected/cc5478a3-096b-4735-9c6e-8ff375821667-kube-api-access-rqgxz\") pod \"dnsmasq-dns-84bb9d8bd9-4n4xr\" (UID: \"cc5478a3-096b-4735-9c6e-8ff375821667\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-4n4xr" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.818676 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd72bd81-8786-47a1-81bd-9ae5d51cc8c2-config\") pod \"dnsmasq-dns-5f854695bc-cjt4j\" (UID: \"dd72bd81-8786-47a1-81bd-9ae5d51cc8c2\") " pod="openstack/dnsmasq-dns-5f854695bc-cjt4j" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.818827 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nwh8r\" (UniqueName: \"kubernetes.io/projected/dd72bd81-8786-47a1-81bd-9ae5d51cc8c2-kube-api-access-nwh8r\") pod \"dnsmasq-dns-5f854695bc-cjt4j\" (UID: \"dd72bd81-8786-47a1-81bd-9ae5d51cc8c2\") " pod="openstack/dnsmasq-dns-5f854695bc-cjt4j" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.818885 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc5478a3-096b-4735-9c6e-8ff375821667-config\") pod \"dnsmasq-dns-84bb9d8bd9-4n4xr\" (UID: \"cc5478a3-096b-4735-9c6e-8ff375821667\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-4n4xr" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.819049 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd72bd81-8786-47a1-81bd-9ae5d51cc8c2-dns-svc\") pod \"dnsmasq-dns-5f854695bc-cjt4j\" (UID: \"dd72bd81-8786-47a1-81bd-9ae5d51cc8c2\") " pod="openstack/dnsmasq-dns-5f854695bc-cjt4j" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.920014 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd72bd81-8786-47a1-81bd-9ae5d51cc8c2-dns-svc\") pod \"dnsmasq-dns-5f854695bc-cjt4j\" (UID: \"dd72bd81-8786-47a1-81bd-9ae5d51cc8c2\") " pod="openstack/dnsmasq-dns-5f854695bc-cjt4j" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.920270 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqgxz\" (UniqueName: \"kubernetes.io/projected/cc5478a3-096b-4735-9c6e-8ff375821667-kube-api-access-rqgxz\") pod \"dnsmasq-dns-84bb9d8bd9-4n4xr\" (UID: \"cc5478a3-096b-4735-9c6e-8ff375821667\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-4n4xr" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.920444 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd72bd81-8786-47a1-81bd-9ae5d51cc8c2-config\") pod \"dnsmasq-dns-5f854695bc-cjt4j\" (UID: \"dd72bd81-8786-47a1-81bd-9ae5d51cc8c2\") " pod="openstack/dnsmasq-dns-5f854695bc-cjt4j" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.920595 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nwh8r\" (UniqueName: \"kubernetes.io/projected/dd72bd81-8786-47a1-81bd-9ae5d51cc8c2-kube-api-access-nwh8r\") pod \"dnsmasq-dns-5f854695bc-cjt4j\" (UID: \"dd72bd81-8786-47a1-81bd-9ae5d51cc8c2\") " pod="openstack/dnsmasq-dns-5f854695bc-cjt4j" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.920728 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc5478a3-096b-4735-9c6e-8ff375821667-config\") pod \"dnsmasq-dns-84bb9d8bd9-4n4xr\" (UID: \"cc5478a3-096b-4735-9c6e-8ff375821667\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-4n4xr" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.921046 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd72bd81-8786-47a1-81bd-9ae5d51cc8c2-dns-svc\") pod \"dnsmasq-dns-5f854695bc-cjt4j\" (UID: \"dd72bd81-8786-47a1-81bd-9ae5d51cc8c2\") " pod="openstack/dnsmasq-dns-5f854695bc-cjt4j" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.921291 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd72bd81-8786-47a1-81bd-9ae5d51cc8c2-config\") pod \"dnsmasq-dns-5f854695bc-cjt4j\" (UID: \"dd72bd81-8786-47a1-81bd-9ae5d51cc8c2\") " pod="openstack/dnsmasq-dns-5f854695bc-cjt4j" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.921440 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc5478a3-096b-4735-9c6e-8ff375821667-config\") pod \"dnsmasq-dns-84bb9d8bd9-4n4xr\" (UID: \"cc5478a3-096b-4735-9c6e-8ff375821667\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-4n4xr" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.939594 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqgxz\" (UniqueName: \"kubernetes.io/projected/cc5478a3-096b-4735-9c6e-8ff375821667-kube-api-access-rqgxz\") pod \"dnsmasq-dns-84bb9d8bd9-4n4xr\" (UID: \"cc5478a3-096b-4735-9c6e-8ff375821667\") " pod="openstack/dnsmasq-dns-84bb9d8bd9-4n4xr" Dec 09 15:47:10 crc kubenswrapper[4894]: I1209 15:47:10.940715 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nwh8r\" (UniqueName: \"kubernetes.io/projected/dd72bd81-8786-47a1-81bd-9ae5d51cc8c2-kube-api-access-nwh8r\") pod \"dnsmasq-dns-5f854695bc-cjt4j\" (UID: \"dd72bd81-8786-47a1-81bd-9ae5d51cc8c2\") " pod="openstack/dnsmasq-dns-5f854695bc-cjt4j" Dec 09 15:47:11 crc kubenswrapper[4894]: I1209 15:47:11.008694 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bb9d8bd9-4n4xr" Dec 09 15:47:11 crc kubenswrapper[4894]: I1209 15:47:11.070759 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f854695bc-cjt4j" Dec 09 15:47:11 crc kubenswrapper[4894]: I1209 15:47:11.500341 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-4n4xr"] Dec 09 15:47:11 crc kubenswrapper[4894]: I1209 15:47:11.508849 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 15:47:11 crc kubenswrapper[4894]: I1209 15:47:11.538810 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84bb9d8bd9-4n4xr" event={"ID":"cc5478a3-096b-4735-9c6e-8ff375821667","Type":"ContainerStarted","Data":"040f09540740fbe90c65dd001e6324418157ab60d9e87cad358c2271adad6eea"} Dec 09 15:47:11 crc kubenswrapper[4894]: I1209 15:47:11.566908 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-cjt4j"] Dec 09 15:47:11 crc kubenswrapper[4894]: W1209 15:47:11.568978 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddd72bd81_8786_47a1_81bd_9ae5d51cc8c2.slice/crio-92462fa45e0ea844173c746844b6b4990804ab18a03da9d8fe46b2123c6be857 WatchSource:0}: Error finding container 92462fa45e0ea844173c746844b6b4990804ab18a03da9d8fe46b2123c6be857: Status 404 returned error can't find the container with id 92462fa45e0ea844173c746844b6b4990804ab18a03da9d8fe46b2123c6be857 Dec 09 15:47:12 crc kubenswrapper[4894]: I1209 15:47:12.258182 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:47:12 crc kubenswrapper[4894]: I1209 15:47:12.258248 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:47:12 crc kubenswrapper[4894]: I1209 15:47:12.258298 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 15:47:12 crc kubenswrapper[4894]: I1209 15:47:12.258937 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3c92ffd29e4b95c91cb30f4577bf2eab4718185324fe428051f5af6db9d8c9a8"} pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 15:47:12 crc kubenswrapper[4894]: I1209 15:47:12.258988 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" containerID="cri-o://3c92ffd29e4b95c91cb30f4577bf2eab4718185324fe428051f5af6db9d8c9a8" gracePeriod=600 Dec 09 15:47:12 crc kubenswrapper[4894]: I1209 15:47:12.546834 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f854695bc-cjt4j" event={"ID":"dd72bd81-8786-47a1-81bd-9ae5d51cc8c2","Type":"ContainerStarted","Data":"92462fa45e0ea844173c746844b6b4990804ab18a03da9d8fe46b2123c6be857"} Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.361731 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-pb2wl" Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.362878 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-pb2wl" Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.483474 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-pb2wl" Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.513833 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-cjt4j"] Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.560779 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-744ffd65bc-2hr9g"] Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.562230 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744ffd65bc-2hr9g" Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.578217 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-744ffd65bc-2hr9g"] Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.588084 4894 generic.go:334] "Generic (PLEG): container finished" podID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerID="3c92ffd29e4b95c91cb30f4577bf2eab4718185324fe428051f5af6db9d8c9a8" exitCode=0 Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.588177 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerDied","Data":"3c92ffd29e4b95c91cb30f4577bf2eab4718185324fe428051f5af6db9d8c9a8"} Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.588240 4894 scope.go:117] "RemoveContainer" containerID="bf0d8a25593f3059a7a7f1286d689a889938309e9429a43b4edb9ebe0cc7eecf" Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.662516 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7qc74\" (UniqueName: \"kubernetes.io/projected/f33e7674-a9a6-4116-807c-9d35ad260b6b-kube-api-access-7qc74\") pod \"dnsmasq-dns-744ffd65bc-2hr9g\" (UID: \"f33e7674-a9a6-4116-807c-9d35ad260b6b\") " pod="openstack/dnsmasq-dns-744ffd65bc-2hr9g" Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.662575 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f33e7674-a9a6-4116-807c-9d35ad260b6b-config\") pod \"dnsmasq-dns-744ffd65bc-2hr9g\" (UID: \"f33e7674-a9a6-4116-807c-9d35ad260b6b\") " pod="openstack/dnsmasq-dns-744ffd65bc-2hr9g" Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.662623 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f33e7674-a9a6-4116-807c-9d35ad260b6b-dns-svc\") pod \"dnsmasq-dns-744ffd65bc-2hr9g\" (UID: \"f33e7674-a9a6-4116-807c-9d35ad260b6b\") " pod="openstack/dnsmasq-dns-744ffd65bc-2hr9g" Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.695370 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-pb2wl" Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.749166 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pb2wl"] Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.764548 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7qc74\" (UniqueName: \"kubernetes.io/projected/f33e7674-a9a6-4116-807c-9d35ad260b6b-kube-api-access-7qc74\") pod \"dnsmasq-dns-744ffd65bc-2hr9g\" (UID: \"f33e7674-a9a6-4116-807c-9d35ad260b6b\") " pod="openstack/dnsmasq-dns-744ffd65bc-2hr9g" Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.764626 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f33e7674-a9a6-4116-807c-9d35ad260b6b-config\") pod \"dnsmasq-dns-744ffd65bc-2hr9g\" (UID: \"f33e7674-a9a6-4116-807c-9d35ad260b6b\") " pod="openstack/dnsmasq-dns-744ffd65bc-2hr9g" Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.764723 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f33e7674-a9a6-4116-807c-9d35ad260b6b-dns-svc\") pod \"dnsmasq-dns-744ffd65bc-2hr9g\" (UID: \"f33e7674-a9a6-4116-807c-9d35ad260b6b\") " pod="openstack/dnsmasq-dns-744ffd65bc-2hr9g" Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.765999 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f33e7674-a9a6-4116-807c-9d35ad260b6b-config\") pod \"dnsmasq-dns-744ffd65bc-2hr9g\" (UID: \"f33e7674-a9a6-4116-807c-9d35ad260b6b\") " pod="openstack/dnsmasq-dns-744ffd65bc-2hr9g" Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.766241 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f33e7674-a9a6-4116-807c-9d35ad260b6b-dns-svc\") pod \"dnsmasq-dns-744ffd65bc-2hr9g\" (UID: \"f33e7674-a9a6-4116-807c-9d35ad260b6b\") " pod="openstack/dnsmasq-dns-744ffd65bc-2hr9g" Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.820831 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7qc74\" (UniqueName: \"kubernetes.io/projected/f33e7674-a9a6-4116-807c-9d35ad260b6b-kube-api-access-7qc74\") pod \"dnsmasq-dns-744ffd65bc-2hr9g\" (UID: \"f33e7674-a9a6-4116-807c-9d35ad260b6b\") " pod="openstack/dnsmasq-dns-744ffd65bc-2hr9g" Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.860544 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-4n4xr"] Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.878149 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-sk7xs"] Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.881699 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-95f5f6995-sk7xs" Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.894783 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-sk7xs"] Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.915335 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744ffd65bc-2hr9g" Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.971392 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb-dns-svc\") pod \"dnsmasq-dns-95f5f6995-sk7xs\" (UID: \"73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb\") " pod="openstack/dnsmasq-dns-95f5f6995-sk7xs" Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.971455 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb-config\") pod \"dnsmasq-dns-95f5f6995-sk7xs\" (UID: \"73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb\") " pod="openstack/dnsmasq-dns-95f5f6995-sk7xs" Dec 09 15:47:13 crc kubenswrapper[4894]: I1209 15:47:13.971502 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbqkj\" (UniqueName: \"kubernetes.io/projected/73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb-kube-api-access-gbqkj\") pod \"dnsmasq-dns-95f5f6995-sk7xs\" (UID: \"73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb\") " pod="openstack/dnsmasq-dns-95f5f6995-sk7xs" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.081420 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb-dns-svc\") pod \"dnsmasq-dns-95f5f6995-sk7xs\" (UID: \"73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb\") " pod="openstack/dnsmasq-dns-95f5f6995-sk7xs" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.081766 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb-config\") pod \"dnsmasq-dns-95f5f6995-sk7xs\" (UID: \"73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb\") " pod="openstack/dnsmasq-dns-95f5f6995-sk7xs" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.081801 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbqkj\" (UniqueName: \"kubernetes.io/projected/73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb-kube-api-access-gbqkj\") pod \"dnsmasq-dns-95f5f6995-sk7xs\" (UID: \"73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb\") " pod="openstack/dnsmasq-dns-95f5f6995-sk7xs" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.083225 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb-dns-svc\") pod \"dnsmasq-dns-95f5f6995-sk7xs\" (UID: \"73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb\") " pod="openstack/dnsmasq-dns-95f5f6995-sk7xs" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.085853 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb-config\") pod \"dnsmasq-dns-95f5f6995-sk7xs\" (UID: \"73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb\") " pod="openstack/dnsmasq-dns-95f5f6995-sk7xs" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.113941 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbqkj\" (UniqueName: \"kubernetes.io/projected/73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb-kube-api-access-gbqkj\") pod \"dnsmasq-dns-95f5f6995-sk7xs\" (UID: \"73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb\") " pod="openstack/dnsmasq-dns-95f5f6995-sk7xs" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.225344 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-95f5f6995-sk7xs" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.506220 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-744ffd65bc-2hr9g"] Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.605155 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerStarted","Data":"06d076fbf334b50ab934fbbefedcfc6bb4d75c33f4047f53bb661bb7b87d68ea"} Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.608662 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-744ffd65bc-2hr9g" event={"ID":"f33e7674-a9a6-4116-807c-9d35ad260b6b","Type":"ContainerStarted","Data":"c048271f44b28d2dcad20b623a726c775219abd8a6f428b2c6593cd9e7a45301"} Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.721466 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.723536 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.725731 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.726101 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.726171 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-6dzck" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.726132 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.726454 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.726467 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.727064 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.731662 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.795602 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-sk7xs"] Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.802511 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.802571 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.802607 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b22pt\" (UniqueName: \"kubernetes.io/projected/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-kube-api-access-b22pt\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.802662 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-server-conf\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.802686 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-pod-info\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.802736 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.803004 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-config-data\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.803025 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.803049 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.803067 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.803089 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.905353 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.905434 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-config-data\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.905458 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.905488 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.905510 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.905543 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.905588 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.905620 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.905675 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b22pt\" (UniqueName: \"kubernetes.io/projected/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-kube-api-access-b22pt\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.905701 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.905712 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-server-conf\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.905739 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-pod-info\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.906492 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.906536 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.907942 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-config-data\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.910054 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.910737 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-server-conf\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.912923 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.916811 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.917312 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.922321 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-pod-info\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.923368 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b22pt\" (UniqueName: \"kubernetes.io/projected/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-kube-api-access-b22pt\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:14 crc kubenswrapper[4894]: I1209 15:47:14.939438 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.039320 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.045189 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.047333 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.048407 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.049277 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.049484 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.049703 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.049893 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-74ccw" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.049981 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.065261 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.072273 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.115763 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.115933 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/44fdaebd-a96c-47f9-a421-a0984f78972e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.116054 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/44fdaebd-a96c-47f9-a421-a0984f78972e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.116097 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.116126 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/44fdaebd-a96c-47f9-a421-a0984f78972e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.116312 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.116499 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/44fdaebd-a96c-47f9-a421-a0984f78972e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.116572 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/44fdaebd-a96c-47f9-a421-a0984f78972e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.116924 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.117117 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dhgw8\" (UniqueName: \"kubernetes.io/projected/44fdaebd-a96c-47f9-a421-a0984f78972e-kube-api-access-dhgw8\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.117157 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.218917 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/44fdaebd-a96c-47f9-a421-a0984f78972e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.219039 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.219080 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/44fdaebd-a96c-47f9-a421-a0984f78972e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.219120 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.219218 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/44fdaebd-a96c-47f9-a421-a0984f78972e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.219267 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/44fdaebd-a96c-47f9-a421-a0984f78972e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.219353 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.219410 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.219442 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dhgw8\" (UniqueName: \"kubernetes.io/projected/44fdaebd-a96c-47f9-a421-a0984f78972e-kube-api-access-dhgw8\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.219551 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.219708 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/44fdaebd-a96c-47f9-a421-a0984f78972e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.225164 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.226028 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.226997 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/44fdaebd-a96c-47f9-a421-a0984f78972e-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.227136 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/44fdaebd-a96c-47f9-a421-a0984f78972e-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.227711 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/44fdaebd-a96c-47f9-a421-a0984f78972e-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.229064 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.230449 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/44fdaebd-a96c-47f9-a421-a0984f78972e-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.231110 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.231280 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.231922 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/44fdaebd-a96c-47f9-a421-a0984f78972e-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.246701 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dhgw8\" (UniqueName: \"kubernetes.io/projected/44fdaebd-a96c-47f9-a421-a0984f78972e-kube-api-access-dhgw8\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.338348 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.394926 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.652044 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-pb2wl" podUID="4cc7356e-a563-4343-9d84-7650b75b3f34" containerName="registry-server" containerID="cri-o://86991a71d70961601eb92e98464d89f445932bd35ad8d61346ff3df108660949" gracePeriod=2 Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.654200 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-95f5f6995-sk7xs" event={"ID":"73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb","Type":"ContainerStarted","Data":"48c29f56bec5cdb814701b023a681e0eecf3e0dd1eb84bdb4e49063ee2c4e43c"} Dec 09 15:47:15 crc kubenswrapper[4894]: I1209 15:47:15.877526 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.095888 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.423281 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.424719 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.432827 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.433119 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-zbtfd" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.433823 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.442050 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.442353 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.448914 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.558421 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d76c34-d0cc-493b-a57a-1923f8d72824-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.558501 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/10d76c34-d0cc-493b-a57a-1923f8d72824-config-data-generated\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.558839 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/10d76c34-d0cc-493b-a57a-1923f8d72824-kolla-config\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.558910 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.558931 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zbfpd\" (UniqueName: \"kubernetes.io/projected/10d76c34-d0cc-493b-a57a-1923f8d72824-kube-api-access-zbfpd\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.558951 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/10d76c34-d0cc-493b-a57a-1923f8d72824-config-data-default\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.558966 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10d76c34-d0cc-493b-a57a-1923f8d72824-operator-scripts\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.559250 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/10d76c34-d0cc-493b-a57a-1923f8d72824-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.660706 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/10d76c34-d0cc-493b-a57a-1923f8d72824-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.660764 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d76c34-d0cc-493b-a57a-1923f8d72824-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.660801 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/10d76c34-d0cc-493b-a57a-1923f8d72824-config-data-generated\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.660833 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/10d76c34-d0cc-493b-a57a-1923f8d72824-kolla-config\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.660860 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zbfpd\" (UniqueName: \"kubernetes.io/projected/10d76c34-d0cc-493b-a57a-1923f8d72824-kube-api-access-zbfpd\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.660876 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.660892 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10d76c34-d0cc-493b-a57a-1923f8d72824-operator-scripts\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.660906 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/10d76c34-d0cc-493b-a57a-1923f8d72824-config-data-default\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.661870 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/10d76c34-d0cc-493b-a57a-1923f8d72824-config-data-default\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.662074 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/10d76c34-d0cc-493b-a57a-1923f8d72824-kolla-config\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.662341 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/10d76c34-d0cc-493b-a57a-1923f8d72824-config-data-generated\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.662417 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.664513 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/10d76c34-d0cc-493b-a57a-1923f8d72824-operator-scripts\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.672177 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"493425e2-b4c7-4a1b-af05-44b4f44bd2b3","Type":"ContainerStarted","Data":"072ef85aa8f529936442b0ea2924d94a9451add40a7d18b7bb424f8bee94fb36"} Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.687979 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/10d76c34-d0cc-493b-a57a-1923f8d72824-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.688744 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10d76c34-d0cc-493b-a57a-1923f8d72824-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.709982 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"44fdaebd-a96c-47f9-a421-a0984f78972e","Type":"ContainerStarted","Data":"988760f4ec9858251ac39f4e2bf3163cd78c8a1ca08415b2fcdb5c44091f3c13"} Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.713650 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zbfpd\" (UniqueName: \"kubernetes.io/projected/10d76c34-d0cc-493b-a57a-1923f8d72824-kube-api-access-zbfpd\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.738725 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-galera-0\" (UID: \"10d76c34-d0cc-493b-a57a-1923f8d72824\") " pod="openstack/openstack-galera-0" Dec 09 15:47:16 crc kubenswrapper[4894]: I1209 15:47:16.758393 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.408492 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.691854 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.693731 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.696335 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.696954 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.697118 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.698925 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-fzkbn" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.723096 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.741083 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"10d76c34-d0cc-493b-a57a-1923f8d72824","Type":"ContainerStarted","Data":"28cc1452fc7978bf31ea884c80feea6521695b83b7ca0c339191ca9dfd9b61ee"} Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.888713 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5caa6181-4755-426f-97fa-50a47d8ea80a-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.888777 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5caa6181-4755-426f-97fa-50a47d8ea80a-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.888797 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-psz7c\" (UniqueName: \"kubernetes.io/projected/5caa6181-4755-426f-97fa-50a47d8ea80a-kube-api-access-psz7c\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.888816 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5caa6181-4755-426f-97fa-50a47d8ea80a-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.888865 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.889444 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5caa6181-4755-426f-97fa-50a47d8ea80a-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.889484 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5caa6181-4755-426f-97fa-50a47d8ea80a-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.889549 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5caa6181-4755-426f-97fa-50a47d8ea80a-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.990390 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5caa6181-4755-426f-97fa-50a47d8ea80a-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.990472 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5caa6181-4755-426f-97fa-50a47d8ea80a-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.990501 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5caa6181-4755-426f-97fa-50a47d8ea80a-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.990522 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-psz7c\" (UniqueName: \"kubernetes.io/projected/5caa6181-4755-426f-97fa-50a47d8ea80a-kube-api-access-psz7c\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.990543 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5caa6181-4755-426f-97fa-50a47d8ea80a-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.990571 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.990764 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5caa6181-4755-426f-97fa-50a47d8ea80a-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.990800 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5caa6181-4755-426f-97fa-50a47d8ea80a-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.990971 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/5caa6181-4755-426f-97fa-50a47d8ea80a-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.991888 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/5caa6181-4755-426f-97fa-50a47d8ea80a-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.992455 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.993206 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/5caa6181-4755-426f-97fa-50a47d8ea80a-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:17 crc kubenswrapper[4894]: I1209 15:47:17.996603 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5caa6181-4755-426f-97fa-50a47d8ea80a-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.000583 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/5caa6181-4755-426f-97fa-50a47d8ea80a-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.001502 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5caa6181-4755-426f-97fa-50a47d8ea80a-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.013793 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-psz7c\" (UniqueName: \"kubernetes.io/projected/5caa6181-4755-426f-97fa-50a47d8ea80a-kube-api-access-psz7c\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.035079 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"openstack-cell1-galera-0\" (UID: \"5caa6181-4755-426f-97fa-50a47d8ea80a\") " pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.048134 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.123075 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.124249 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.129764 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-gxwf7" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.129999 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.130504 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.135808 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.300690 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb\") " pod="openstack/memcached-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.301512 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb\") " pod="openstack/memcached-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.301552 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46smk\" (UniqueName: \"kubernetes.io/projected/9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb-kube-api-access-46smk\") pod \"memcached-0\" (UID: \"9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb\") " pod="openstack/memcached-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.301581 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb-kolla-config\") pod \"memcached-0\" (UID: \"9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb\") " pod="openstack/memcached-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.301601 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb-config-data\") pod \"memcached-0\" (UID: \"9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb\") " pod="openstack/memcached-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.404436 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb\") " pod="openstack/memcached-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.404509 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb\") " pod="openstack/memcached-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.404579 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46smk\" (UniqueName: \"kubernetes.io/projected/9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb-kube-api-access-46smk\") pod \"memcached-0\" (UID: \"9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb\") " pod="openstack/memcached-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.404612 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb-kolla-config\") pod \"memcached-0\" (UID: \"9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb\") " pod="openstack/memcached-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.404653 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb-config-data\") pod \"memcached-0\" (UID: \"9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb\") " pod="openstack/memcached-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.408132 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb-config-data\") pod \"memcached-0\" (UID: \"9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb\") " pod="openstack/memcached-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.409047 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb-kolla-config\") pod \"memcached-0\" (UID: \"9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb\") " pod="openstack/memcached-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.415570 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb-combined-ca-bundle\") pod \"memcached-0\" (UID: \"9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb\") " pod="openstack/memcached-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.423569 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb-memcached-tls-certs\") pod \"memcached-0\" (UID: \"9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb\") " pod="openstack/memcached-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.443907 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46smk\" (UniqueName: \"kubernetes.io/projected/9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb-kube-api-access-46smk\") pod \"memcached-0\" (UID: \"9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb\") " pod="openstack/memcached-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.457765 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.615828 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.748827 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"5caa6181-4755-426f-97fa-50a47d8ea80a","Type":"ContainerStarted","Data":"03adb282291e4974bf68d306282d037f0290470f8141ccef8432ef3cdb94b2a4"} Dec 09 15:47:18 crc kubenswrapper[4894]: I1209 15:47:18.962225 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Dec 09 15:47:18 crc kubenswrapper[4894]: W1209 15:47:18.985850 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9ef2f4b3_b303_4b17_b4e8_fc2b8929bdfb.slice/crio-e4701ce63b0c69c4c11f18c72a5a42ac9c253648dc89ea15037793a3b2493047 WatchSource:0}: Error finding container e4701ce63b0c69c4c11f18c72a5a42ac9c253648dc89ea15037793a3b2493047: Status 404 returned error can't find the container with id e4701ce63b0c69c4c11f18c72a5a42ac9c253648dc89ea15037793a3b2493047 Dec 09 15:47:19 crc kubenswrapper[4894]: I1209 15:47:19.797502 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb","Type":"ContainerStarted","Data":"e4701ce63b0c69c4c11f18c72a5a42ac9c253648dc89ea15037793a3b2493047"} Dec 09 15:47:19 crc kubenswrapper[4894]: I1209 15:47:19.801775 4894 generic.go:334] "Generic (PLEG): container finished" podID="4cc7356e-a563-4343-9d84-7650b75b3f34" containerID="86991a71d70961601eb92e98464d89f445932bd35ad8d61346ff3df108660949" exitCode=0 Dec 09 15:47:19 crc kubenswrapper[4894]: I1209 15:47:19.801817 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pb2wl" event={"ID":"4cc7356e-a563-4343-9d84-7650b75b3f34","Type":"ContainerDied","Data":"86991a71d70961601eb92e98464d89f445932bd35ad8d61346ff3df108660949"} Dec 09 15:47:20 crc kubenswrapper[4894]: I1209 15:47:20.082140 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 15:47:20 crc kubenswrapper[4894]: I1209 15:47:20.083831 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 15:47:20 crc kubenswrapper[4894]: I1209 15:47:20.086890 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-vnzhh" Dec 09 15:47:20 crc kubenswrapper[4894]: I1209 15:47:20.143221 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 15:47:20 crc kubenswrapper[4894]: I1209 15:47:20.155577 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gk2n5\" (UniqueName: \"kubernetes.io/projected/48d401c8-b930-410e-a313-80399a664641-kube-api-access-gk2n5\") pod \"kube-state-metrics-0\" (UID: \"48d401c8-b930-410e-a313-80399a664641\") " pod="openstack/kube-state-metrics-0" Dec 09 15:47:20 crc kubenswrapper[4894]: I1209 15:47:20.257932 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gk2n5\" (UniqueName: \"kubernetes.io/projected/48d401c8-b930-410e-a313-80399a664641-kube-api-access-gk2n5\") pod \"kube-state-metrics-0\" (UID: \"48d401c8-b930-410e-a313-80399a664641\") " pod="openstack/kube-state-metrics-0" Dec 09 15:47:20 crc kubenswrapper[4894]: I1209 15:47:20.293046 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gk2n5\" (UniqueName: \"kubernetes.io/projected/48d401c8-b930-410e-a313-80399a664641-kube-api-access-gk2n5\") pod \"kube-state-metrics-0\" (UID: \"48d401c8-b930-410e-a313-80399a664641\") " pod="openstack/kube-state-metrics-0" Dec 09 15:47:20 crc kubenswrapper[4894]: I1209 15:47:20.430926 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 15:47:23 crc kubenswrapper[4894]: E1209 15:47:23.359703 4894 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 86991a71d70961601eb92e98464d89f445932bd35ad8d61346ff3df108660949 is running failed: container process not found" containerID="86991a71d70961601eb92e98464d89f445932bd35ad8d61346ff3df108660949" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 15:47:23 crc kubenswrapper[4894]: E1209 15:47:23.360333 4894 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 86991a71d70961601eb92e98464d89f445932bd35ad8d61346ff3df108660949 is running failed: container process not found" containerID="86991a71d70961601eb92e98464d89f445932bd35ad8d61346ff3df108660949" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 15:47:23 crc kubenswrapper[4894]: E1209 15:47:23.360571 4894 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 86991a71d70961601eb92e98464d89f445932bd35ad8d61346ff3df108660949 is running failed: container process not found" containerID="86991a71d70961601eb92e98464d89f445932bd35ad8d61346ff3df108660949" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 15:47:23 crc kubenswrapper[4894]: E1209 15:47:23.360608 4894 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 86991a71d70961601eb92e98464d89f445932bd35ad8d61346ff3df108660949 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-pb2wl" podUID="4cc7356e-a563-4343-9d84-7650b75b3f34" containerName="registry-server" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.270703 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-vk5h6"] Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.272080 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.276969 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-z7vcg"] Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.278480 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.283363 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.283658 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-8r749" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.283781 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.294187 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vk5h6"] Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.308120 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-z7vcg"] Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.333936 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf53d7df-bcc4-46b3-b874-dac48eca5bc4-combined-ca-bundle\") pod \"ovn-controller-vk5h6\" (UID: \"cf53d7df-bcc4-46b3-b874-dac48eca5bc4\") " pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.334001 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/b0fedf20-9f17-4d49-bef9-c9476a5027cf-etc-ovs\") pod \"ovn-controller-ovs-z7vcg\" (UID: \"b0fedf20-9f17-4d49-bef9-c9476a5027cf\") " pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.334094 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf53d7df-bcc4-46b3-b874-dac48eca5bc4-scripts\") pod \"ovn-controller-vk5h6\" (UID: \"cf53d7df-bcc4-46b3-b874-dac48eca5bc4\") " pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.334130 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0fedf20-9f17-4d49-bef9-c9476a5027cf-scripts\") pod \"ovn-controller-ovs-z7vcg\" (UID: \"b0fedf20-9f17-4d49-bef9-c9476a5027cf\") " pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.334173 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cf53d7df-bcc4-46b3-b874-dac48eca5bc4-var-log-ovn\") pod \"ovn-controller-vk5h6\" (UID: \"cf53d7df-bcc4-46b3-b874-dac48eca5bc4\") " pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.334195 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf53d7df-bcc4-46b3-b874-dac48eca5bc4-ovn-controller-tls-certs\") pod \"ovn-controller-vk5h6\" (UID: \"cf53d7df-bcc4-46b3-b874-dac48eca5bc4\") " pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.334236 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b0fedf20-9f17-4d49-bef9-c9476a5027cf-var-run\") pod \"ovn-controller-ovs-z7vcg\" (UID: \"b0fedf20-9f17-4d49-bef9-c9476a5027cf\") " pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.334257 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b0fedf20-9f17-4d49-bef9-c9476a5027cf-var-log\") pod \"ovn-controller-ovs-z7vcg\" (UID: \"b0fedf20-9f17-4d49-bef9-c9476a5027cf\") " pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.334278 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlsjk\" (UniqueName: \"kubernetes.io/projected/b0fedf20-9f17-4d49-bef9-c9476a5027cf-kube-api-access-xlsjk\") pod \"ovn-controller-ovs-z7vcg\" (UID: \"b0fedf20-9f17-4d49-bef9-c9476a5027cf\") " pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.334360 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cf53d7df-bcc4-46b3-b874-dac48eca5bc4-var-run-ovn\") pod \"ovn-controller-vk5h6\" (UID: \"cf53d7df-bcc4-46b3-b874-dac48eca5bc4\") " pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.334388 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/b0fedf20-9f17-4d49-bef9-c9476a5027cf-var-lib\") pod \"ovn-controller-ovs-z7vcg\" (UID: \"b0fedf20-9f17-4d49-bef9-c9476a5027cf\") " pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.334433 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tzld7\" (UniqueName: \"kubernetes.io/projected/cf53d7df-bcc4-46b3-b874-dac48eca5bc4-kube-api-access-tzld7\") pod \"ovn-controller-vk5h6\" (UID: \"cf53d7df-bcc4-46b3-b874-dac48eca5bc4\") " pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.334488 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cf53d7df-bcc4-46b3-b874-dac48eca5bc4-var-run\") pod \"ovn-controller-vk5h6\" (UID: \"cf53d7df-bcc4-46b3-b874-dac48eca5bc4\") " pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.436246 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b0fedf20-9f17-4d49-bef9-c9476a5027cf-var-run\") pod \"ovn-controller-ovs-z7vcg\" (UID: \"b0fedf20-9f17-4d49-bef9-c9476a5027cf\") " pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.436294 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b0fedf20-9f17-4d49-bef9-c9476a5027cf-var-log\") pod \"ovn-controller-ovs-z7vcg\" (UID: \"b0fedf20-9f17-4d49-bef9-c9476a5027cf\") " pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.436317 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlsjk\" (UniqueName: \"kubernetes.io/projected/b0fedf20-9f17-4d49-bef9-c9476a5027cf-kube-api-access-xlsjk\") pod \"ovn-controller-ovs-z7vcg\" (UID: \"b0fedf20-9f17-4d49-bef9-c9476a5027cf\") " pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.436386 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cf53d7df-bcc4-46b3-b874-dac48eca5bc4-var-run-ovn\") pod \"ovn-controller-vk5h6\" (UID: \"cf53d7df-bcc4-46b3-b874-dac48eca5bc4\") " pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.436414 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/b0fedf20-9f17-4d49-bef9-c9476a5027cf-var-lib\") pod \"ovn-controller-ovs-z7vcg\" (UID: \"b0fedf20-9f17-4d49-bef9-c9476a5027cf\") " pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.436461 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tzld7\" (UniqueName: \"kubernetes.io/projected/cf53d7df-bcc4-46b3-b874-dac48eca5bc4-kube-api-access-tzld7\") pod \"ovn-controller-vk5h6\" (UID: \"cf53d7df-bcc4-46b3-b874-dac48eca5bc4\") " pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.436498 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cf53d7df-bcc4-46b3-b874-dac48eca5bc4-var-run\") pod \"ovn-controller-vk5h6\" (UID: \"cf53d7df-bcc4-46b3-b874-dac48eca5bc4\") " pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.436534 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf53d7df-bcc4-46b3-b874-dac48eca5bc4-combined-ca-bundle\") pod \"ovn-controller-vk5h6\" (UID: \"cf53d7df-bcc4-46b3-b874-dac48eca5bc4\") " pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.436559 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/b0fedf20-9f17-4d49-bef9-c9476a5027cf-etc-ovs\") pod \"ovn-controller-ovs-z7vcg\" (UID: \"b0fedf20-9f17-4d49-bef9-c9476a5027cf\") " pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.436585 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf53d7df-bcc4-46b3-b874-dac48eca5bc4-scripts\") pod \"ovn-controller-vk5h6\" (UID: \"cf53d7df-bcc4-46b3-b874-dac48eca5bc4\") " pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.436607 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0fedf20-9f17-4d49-bef9-c9476a5027cf-scripts\") pod \"ovn-controller-ovs-z7vcg\" (UID: \"b0fedf20-9f17-4d49-bef9-c9476a5027cf\") " pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.436658 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cf53d7df-bcc4-46b3-b874-dac48eca5bc4-var-log-ovn\") pod \"ovn-controller-vk5h6\" (UID: \"cf53d7df-bcc4-46b3-b874-dac48eca5bc4\") " pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.436680 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf53d7df-bcc4-46b3-b874-dac48eca5bc4-ovn-controller-tls-certs\") pod \"ovn-controller-vk5h6\" (UID: \"cf53d7df-bcc4-46b3-b874-dac48eca5bc4\") " pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.436753 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/b0fedf20-9f17-4d49-bef9-c9476a5027cf-var-run\") pod \"ovn-controller-ovs-z7vcg\" (UID: \"b0fedf20-9f17-4d49-bef9-c9476a5027cf\") " pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.436833 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/cf53d7df-bcc4-46b3-b874-dac48eca5bc4-var-run\") pod \"ovn-controller-vk5h6\" (UID: \"cf53d7df-bcc4-46b3-b874-dac48eca5bc4\") " pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.436914 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/b0fedf20-9f17-4d49-bef9-c9476a5027cf-var-log\") pod \"ovn-controller-ovs-z7vcg\" (UID: \"b0fedf20-9f17-4d49-bef9-c9476a5027cf\") " pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.437143 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/cf53d7df-bcc4-46b3-b874-dac48eca5bc4-var-run-ovn\") pod \"ovn-controller-vk5h6\" (UID: \"cf53d7df-bcc4-46b3-b874-dac48eca5bc4\") " pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.437372 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/b0fedf20-9f17-4d49-bef9-c9476a5027cf-var-lib\") pod \"ovn-controller-ovs-z7vcg\" (UID: \"b0fedf20-9f17-4d49-bef9-c9476a5027cf\") " pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.437536 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/cf53d7df-bcc4-46b3-b874-dac48eca5bc4-var-log-ovn\") pod \"ovn-controller-vk5h6\" (UID: \"cf53d7df-bcc4-46b3-b874-dac48eca5bc4\") " pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.437775 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/b0fedf20-9f17-4d49-bef9-c9476a5027cf-etc-ovs\") pod \"ovn-controller-ovs-z7vcg\" (UID: \"b0fedf20-9f17-4d49-bef9-c9476a5027cf\") " pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.439396 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/b0fedf20-9f17-4d49-bef9-c9476a5027cf-scripts\") pod \"ovn-controller-ovs-z7vcg\" (UID: \"b0fedf20-9f17-4d49-bef9-c9476a5027cf\") " pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.440384 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/cf53d7df-bcc4-46b3-b874-dac48eca5bc4-scripts\") pod \"ovn-controller-vk5h6\" (UID: \"cf53d7df-bcc4-46b3-b874-dac48eca5bc4\") " pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.443887 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cf53d7df-bcc4-46b3-b874-dac48eca5bc4-combined-ca-bundle\") pod \"ovn-controller-vk5h6\" (UID: \"cf53d7df-bcc4-46b3-b874-dac48eca5bc4\") " pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.446064 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/cf53d7df-bcc4-46b3-b874-dac48eca5bc4-ovn-controller-tls-certs\") pod \"ovn-controller-vk5h6\" (UID: \"cf53d7df-bcc4-46b3-b874-dac48eca5bc4\") " pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.454011 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlsjk\" (UniqueName: \"kubernetes.io/projected/b0fedf20-9f17-4d49-bef9-c9476a5027cf-kube-api-access-xlsjk\") pod \"ovn-controller-ovs-z7vcg\" (UID: \"b0fedf20-9f17-4d49-bef9-c9476a5027cf\") " pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.469986 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tzld7\" (UniqueName: \"kubernetes.io/projected/cf53d7df-bcc4-46b3-b874-dac48eca5bc4-kube-api-access-tzld7\") pod \"ovn-controller-vk5h6\" (UID: \"cf53d7df-bcc4-46b3-b874-dac48eca5bc4\") " pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.605258 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:24 crc kubenswrapper[4894]: I1209 15:47:24.611988 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:47:26 crc kubenswrapper[4894]: I1209 15:47:26.104455 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pb2wl" Dec 09 15:47:26 crc kubenswrapper[4894]: I1209 15:47:26.169372 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cc7356e-a563-4343-9d84-7650b75b3f34-utilities\") pod \"4cc7356e-a563-4343-9d84-7650b75b3f34\" (UID: \"4cc7356e-a563-4343-9d84-7650b75b3f34\") " Dec 09 15:47:26 crc kubenswrapper[4894]: I1209 15:47:26.169555 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cc7356e-a563-4343-9d84-7650b75b3f34-catalog-content\") pod \"4cc7356e-a563-4343-9d84-7650b75b3f34\" (UID: \"4cc7356e-a563-4343-9d84-7650b75b3f34\") " Dec 09 15:47:26 crc kubenswrapper[4894]: I1209 15:47:26.169591 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5c5sl\" (UniqueName: \"kubernetes.io/projected/4cc7356e-a563-4343-9d84-7650b75b3f34-kube-api-access-5c5sl\") pod \"4cc7356e-a563-4343-9d84-7650b75b3f34\" (UID: \"4cc7356e-a563-4343-9d84-7650b75b3f34\") " Dec 09 15:47:26 crc kubenswrapper[4894]: I1209 15:47:26.179894 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cc7356e-a563-4343-9d84-7650b75b3f34-utilities" (OuterVolumeSpecName: "utilities") pod "4cc7356e-a563-4343-9d84-7650b75b3f34" (UID: "4cc7356e-a563-4343-9d84-7650b75b3f34"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:47:26 crc kubenswrapper[4894]: I1209 15:47:26.201856 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cc7356e-a563-4343-9d84-7650b75b3f34-kube-api-access-5c5sl" (OuterVolumeSpecName: "kube-api-access-5c5sl") pod "4cc7356e-a563-4343-9d84-7650b75b3f34" (UID: "4cc7356e-a563-4343-9d84-7650b75b3f34"). InnerVolumeSpecName "kube-api-access-5c5sl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:47:26 crc kubenswrapper[4894]: I1209 15:47:26.203588 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cc7356e-a563-4343-9d84-7650b75b3f34-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "4cc7356e-a563-4343-9d84-7650b75b3f34" (UID: "4cc7356e-a563-4343-9d84-7650b75b3f34"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:47:26 crc kubenswrapper[4894]: I1209 15:47:26.272252 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4cc7356e-a563-4343-9d84-7650b75b3f34-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 15:47:26 crc kubenswrapper[4894]: I1209 15:47:26.272289 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5c5sl\" (UniqueName: \"kubernetes.io/projected/4cc7356e-a563-4343-9d84-7650b75b3f34-kube-api-access-5c5sl\") on node \"crc\" DevicePath \"\"" Dec 09 15:47:26 crc kubenswrapper[4894]: I1209 15:47:26.272301 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4cc7356e-a563-4343-9d84-7650b75b3f34-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 15:47:26 crc kubenswrapper[4894]: I1209 15:47:26.869332 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-pb2wl" event={"ID":"4cc7356e-a563-4343-9d84-7650b75b3f34","Type":"ContainerDied","Data":"df0bb7e8fc52aad5edc742dfdc431bd6ada55dcda8c85ebc3ada004e83a40685"} Dec 09 15:47:26 crc kubenswrapper[4894]: I1209 15:47:26.869383 4894 scope.go:117] "RemoveContainer" containerID="86991a71d70961601eb92e98464d89f445932bd35ad8d61346ff3df108660949" Dec 09 15:47:26 crc kubenswrapper[4894]: I1209 15:47:26.869471 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-pb2wl" Dec 09 15:47:26 crc kubenswrapper[4894]: I1209 15:47:26.913871 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-pb2wl"] Dec 09 15:47:26 crc kubenswrapper[4894]: I1209 15:47:26.919323 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-pb2wl"] Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.321734 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 09 15:47:27 crc kubenswrapper[4894]: E1209 15:47:27.322346 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cc7356e-a563-4343-9d84-7650b75b3f34" containerName="registry-server" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.322367 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cc7356e-a563-4343-9d84-7650b75b3f34" containerName="registry-server" Dec 09 15:47:27 crc kubenswrapper[4894]: E1209 15:47:27.322397 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cc7356e-a563-4343-9d84-7650b75b3f34" containerName="extract-content" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.322406 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cc7356e-a563-4343-9d84-7650b75b3f34" containerName="extract-content" Dec 09 15:47:27 crc kubenswrapper[4894]: E1209 15:47:27.322434 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cc7356e-a563-4343-9d84-7650b75b3f34" containerName="extract-utilities" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.322441 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cc7356e-a563-4343-9d84-7650b75b3f34" containerName="extract-utilities" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.334408 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cc7356e-a563-4343-9d84-7650b75b3f34" containerName="registry-server" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.335611 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.335777 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.340208 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.342285 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-h586w" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.342340 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.342563 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.344032 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.401753 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.402131 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6e0e621-ea8e-4256-8c04-74c54686708a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.402172 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5dbxj\" (UniqueName: \"kubernetes.io/projected/a6e0e621-ea8e-4256-8c04-74c54686708a-kube-api-access-5dbxj\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.402206 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a6e0e621-ea8e-4256-8c04-74c54686708a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.402259 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6e0e621-ea8e-4256-8c04-74c54686708a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.402289 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6e0e621-ea8e-4256-8c04-74c54686708a-config\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.402351 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6e0e621-ea8e-4256-8c04-74c54686708a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.402407 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6e0e621-ea8e-4256-8c04-74c54686708a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.504077 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.504129 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6e0e621-ea8e-4256-8c04-74c54686708a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.504152 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5dbxj\" (UniqueName: \"kubernetes.io/projected/a6e0e621-ea8e-4256-8c04-74c54686708a-kube-api-access-5dbxj\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.504297 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a6e0e621-ea8e-4256-8c04-74c54686708a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.504320 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6e0e621-ea8e-4256-8c04-74c54686708a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.504347 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6e0e621-ea8e-4256-8c04-74c54686708a-config\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.504380 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6e0e621-ea8e-4256-8c04-74c54686708a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.504401 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6e0e621-ea8e-4256-8c04-74c54686708a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.504567 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.505967 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6e0e621-ea8e-4256-8c04-74c54686708a-config\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.506003 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/a6e0e621-ea8e-4256-8c04-74c54686708a-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.507293 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/a6e0e621-ea8e-4256-8c04-74c54686708a-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.511128 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6e0e621-ea8e-4256-8c04-74c54686708a-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.513476 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a6e0e621-ea8e-4256-8c04-74c54686708a-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.527846 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/a6e0e621-ea8e-4256-8c04-74c54686708a-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.530111 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.531708 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.533739 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5dbxj\" (UniqueName: \"kubernetes.io/projected/a6e0e621-ea8e-4256-8c04-74c54686708a-kube-api-access-5dbxj\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.535505 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-8bl7d" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.535992 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.536099 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.536336 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.546326 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.557068 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"ovsdbserver-nb-0\" (UID: \"a6e0e621-ea8e-4256-8c04-74c54686708a\") " pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.605368 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4eebb217-538f-4f1a-a4f8-93371486803d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.605419 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.605478 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmm9h\" (UniqueName: \"kubernetes.io/projected/4eebb217-538f-4f1a-a4f8-93371486803d-kube-api-access-nmm9h\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.605514 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4eebb217-538f-4f1a-a4f8-93371486803d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.605573 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4eebb217-538f-4f1a-a4f8-93371486803d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.605658 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eebb217-538f-4f1a-a4f8-93371486803d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.605692 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4eebb217-538f-4f1a-a4f8-93371486803d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.606264 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4eebb217-538f-4f1a-a4f8-93371486803d-config\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.667379 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.707343 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4eebb217-538f-4f1a-a4f8-93371486803d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.707385 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.707453 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmm9h\" (UniqueName: \"kubernetes.io/projected/4eebb217-538f-4f1a-a4f8-93371486803d-kube-api-access-nmm9h\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.707486 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4eebb217-538f-4f1a-a4f8-93371486803d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.707521 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4eebb217-538f-4f1a-a4f8-93371486803d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.707552 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eebb217-538f-4f1a-a4f8-93371486803d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.707596 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4eebb217-538f-4f1a-a4f8-93371486803d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.707615 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4eebb217-538f-4f1a-a4f8-93371486803d-config\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.707616 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.708726 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4eebb217-538f-4f1a-a4f8-93371486803d-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.711834 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/4eebb217-538f-4f1a-a4f8-93371486803d-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.713377 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4eebb217-538f-4f1a-a4f8-93371486803d-config\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.715452 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/4eebb217-538f-4f1a-a4f8-93371486803d-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.716546 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/4eebb217-538f-4f1a-a4f8-93371486803d-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.726858 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4eebb217-538f-4f1a-a4f8-93371486803d-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.729188 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmm9h\" (UniqueName: \"kubernetes.io/projected/4eebb217-538f-4f1a-a4f8-93371486803d-kube-api-access-nmm9h\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.733593 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"ovsdbserver-sb-0\" (UID: \"4eebb217-538f-4f1a-a4f8-93371486803d\") " pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:27 crc kubenswrapper[4894]: I1209 15:47:27.910166 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:28 crc kubenswrapper[4894]: I1209 15:47:28.155655 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cc7356e-a563-4343-9d84-7650b75b3f34" path="/var/lib/kubelet/pods/4cc7356e-a563-4343-9d84-7650b75b3f34/volumes" Dec 09 15:47:34 crc kubenswrapper[4894]: I1209 15:47:34.461126 4894 scope.go:117] "RemoveContainer" containerID="6410ddca9172568afb45375e59361b3df8dba0ccc28566f3b35a14f06ca50fcf" Dec 09 15:47:43 crc kubenswrapper[4894]: I1209 15:47:43.067777 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 15:47:43 crc kubenswrapper[4894]: I1209 15:47:43.274834 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-z7vcg"] Dec 09 15:47:44 crc kubenswrapper[4894]: I1209 15:47:44.227168 4894 scope.go:117] "RemoveContainer" containerID="d367cf4408e633f7ad5215745e14cf7c63996a9f4fdafdfdecbd17b904878809" Dec 09 15:47:44 crc kubenswrapper[4894]: E1209 15:47:44.254660 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33" Dec 09 15:47:44 crc kubenswrapper[4894]: E1209 15:47:44.254836 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-nwh8r,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-5f854695bc-cjt4j_openstack(dd72bd81-8786-47a1-81bd-9ae5d51cc8c2): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 15:47:44 crc kubenswrapper[4894]: E1209 15:47:44.256229 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-5f854695bc-cjt4j" podUID="dd72bd81-8786-47a1-81bd-9ae5d51cc8c2" Dec 09 15:47:44 crc kubenswrapper[4894]: E1209 15:47:44.265872 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33" Dec 09 15:47:44 crc kubenswrapper[4894]: E1209 15:47:44.265988 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n68chd6h679hbfh55fhc6h5ffh5d8h94h56ch589hb4hc5h57bh677hcdh655h8dh667h675h654h66ch567h8fh659h5b4h675h566h55bh54h67dh6dq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7qc74,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-744ffd65bc-2hr9g_openstack(f33e7674-a9a6-4116-807c-9d35ad260b6b): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 15:47:44 crc kubenswrapper[4894]: E1209 15:47:44.267491 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-744ffd65bc-2hr9g" podUID="f33e7674-a9a6-4116-807c-9d35ad260b6b" Dec 09 15:47:44 crc kubenswrapper[4894]: E1209 15:47:44.292884 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33" Dec 09 15:47:44 crc kubenswrapper[4894]: E1209 15:47:44.293076 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n659h4h664hbh658h587h67ch89h587h8fh679hc6hf9h55fh644h5d5h698h68dh5cdh5ffh669h54ch9h689hb8hd4h5bfhd8h5d7h5fh665h574q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gbqkj,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-95f5f6995-sk7xs_openstack(73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 15:47:44 crc kubenswrapper[4894]: E1209 15:47:44.294585 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-95f5f6995-sk7xs" podUID="73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb" Dec 09 15:47:44 crc kubenswrapper[4894]: E1209 15:47:44.325804 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33" Dec 09 15:47:44 crc kubenswrapper[4894]: E1209 15:47:44.326166 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rqgxz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-84bb9d8bd9-4n4xr_openstack(cc5478a3-096b-4735-9c6e-8ff375821667): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 15:47:44 crc kubenswrapper[4894]: E1209 15:47:44.327499 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-84bb9d8bd9-4n4xr" podUID="cc5478a3-096b-4735-9c6e-8ff375821667" Dec 09 15:47:44 crc kubenswrapper[4894]: I1209 15:47:44.692346 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Dec 09 15:47:44 crc kubenswrapper[4894]: I1209 15:47:44.701790 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vk5h6"] Dec 09 15:47:44 crc kubenswrapper[4894]: W1209 15:47:44.705599 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda6e0e621_ea8e_4256_8c04_74c54686708a.slice/crio-5cfe37add762de54c2730f3bb1b8ef3c94191f330eba03aa69d4577b38a88714 WatchSource:0}: Error finding container 5cfe37add762de54c2730f3bb1b8ef3c94191f330eba03aa69d4577b38a88714: Status 404 returned error can't find the container with id 5cfe37add762de54c2730f3bb1b8ef3c94191f330eba03aa69d4577b38a88714 Dec 09 15:47:44 crc kubenswrapper[4894]: W1209 15:47:44.706667 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podcf53d7df_bcc4_46b3_b874_dac48eca5bc4.slice/crio-cd0e4bee94c9e4ab989cb934552f8a24d29e169f53b39a9406918f3abfa12b95 WatchSource:0}: Error finding container cd0e4bee94c9e4ab989cb934552f8a24d29e169f53b39a9406918f3abfa12b95: Status 404 returned error can't find the container with id cd0e4bee94c9e4ab989cb934552f8a24d29e169f53b39a9406918f3abfa12b95 Dec 09 15:47:44 crc kubenswrapper[4894]: I1209 15:47:44.963970 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.018507 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"48d401c8-b930-410e-a313-80399a664641","Type":"ContainerStarted","Data":"afd9f3c9cc738ec79ce7df87c0ecd6b32a700e7b65cde148fe85b718b531f086"} Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.019670 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vk5h6" event={"ID":"cf53d7df-bcc4-46b3-b874-dac48eca5bc4","Type":"ContainerStarted","Data":"cd0e4bee94c9e4ab989cb934552f8a24d29e169f53b39a9406918f3abfa12b95"} Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.021127 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb","Type":"ContainerStarted","Data":"ed8c08cd3e3b9a439b0c6e5989f533c721a00b7a1348aa335093f49aab68d983"} Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.021247 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.022454 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-z7vcg" event={"ID":"b0fedf20-9f17-4d49-bef9-c9476a5027cf","Type":"ContainerStarted","Data":"4c71e9748041cc322558f5a1450eeb03007b2521d23bdde7943741b85606c0a2"} Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.024661 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"10d76c34-d0cc-493b-a57a-1923f8d72824","Type":"ContainerStarted","Data":"9485aa05471a006a1012a9a84b30857454c22e3295643dc269dd2ea3878d44cc"} Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.029409 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a6e0e621-ea8e-4256-8c04-74c54686708a","Type":"ContainerStarted","Data":"5cfe37add762de54c2730f3bb1b8ef3c94191f330eba03aa69d4577b38a88714"} Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.031456 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"5caa6181-4755-426f-97fa-50a47d8ea80a","Type":"ContainerStarted","Data":"565ae530aed1596c22f18435e7871b4ab6e95db11efb296135dfe54ce89a5d75"} Dec 09 15:47:45 crc kubenswrapper[4894]: E1209 15:47:45.033723 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33\\\"\"" pod="openstack/dnsmasq-dns-744ffd65bc-2hr9g" podUID="f33e7674-a9a6-4116-807c-9d35ad260b6b" Dec 09 15:47:45 crc kubenswrapper[4894]: E1209 15:47:45.034512 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-neutron-server@sha256:ea0bf67f1aa5d95a9a07b9c8692c293470f1311792c55d3d57f1f92e56689c33\\\"\"" pod="openstack/dnsmasq-dns-95f5f6995-sk7xs" podUID="73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb" Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.042892 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=1.7797713960000001 podStartE2EDuration="27.042865705s" podCreationTimestamp="2025-12-09 15:47:18 +0000 UTC" firstStartedPulling="2025-12-09 15:47:18.990373901 +0000 UTC m=+933.309584570" lastFinishedPulling="2025-12-09 15:47:44.25346821 +0000 UTC m=+958.572678879" observedRunningTime="2025-12-09 15:47:45.038551849 +0000 UTC m=+959.357762528" watchObservedRunningTime="2025-12-09 15:47:45.042865705 +0000 UTC m=+959.362076374" Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.700184 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f854695bc-cjt4j" Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.707777 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bb9d8bd9-4n4xr" Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.749085 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd72bd81-8786-47a1-81bd-9ae5d51cc8c2-config\") pod \"dd72bd81-8786-47a1-81bd-9ae5d51cc8c2\" (UID: \"dd72bd81-8786-47a1-81bd-9ae5d51cc8c2\") " Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.749172 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd72bd81-8786-47a1-81bd-9ae5d51cc8c2-dns-svc\") pod \"dd72bd81-8786-47a1-81bd-9ae5d51cc8c2\" (UID: \"dd72bd81-8786-47a1-81bd-9ae5d51cc8c2\") " Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.749242 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nwh8r\" (UniqueName: \"kubernetes.io/projected/dd72bd81-8786-47a1-81bd-9ae5d51cc8c2-kube-api-access-nwh8r\") pod \"dd72bd81-8786-47a1-81bd-9ae5d51cc8c2\" (UID: \"dd72bd81-8786-47a1-81bd-9ae5d51cc8c2\") " Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.749864 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd72bd81-8786-47a1-81bd-9ae5d51cc8c2-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dd72bd81-8786-47a1-81bd-9ae5d51cc8c2" (UID: "dd72bd81-8786-47a1-81bd-9ae5d51cc8c2"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.749902 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dd72bd81-8786-47a1-81bd-9ae5d51cc8c2-config" (OuterVolumeSpecName: "config") pod "dd72bd81-8786-47a1-81bd-9ae5d51cc8c2" (UID: "dd72bd81-8786-47a1-81bd-9ae5d51cc8c2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.755882 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd72bd81-8786-47a1-81bd-9ae5d51cc8c2-kube-api-access-nwh8r" (OuterVolumeSpecName: "kube-api-access-nwh8r") pod "dd72bd81-8786-47a1-81bd-9ae5d51cc8c2" (UID: "dd72bd81-8786-47a1-81bd-9ae5d51cc8c2"). InnerVolumeSpecName "kube-api-access-nwh8r". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.850392 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rqgxz\" (UniqueName: \"kubernetes.io/projected/cc5478a3-096b-4735-9c6e-8ff375821667-kube-api-access-rqgxz\") pod \"cc5478a3-096b-4735-9c6e-8ff375821667\" (UID: \"cc5478a3-096b-4735-9c6e-8ff375821667\") " Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.850629 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc5478a3-096b-4735-9c6e-8ff375821667-config\") pod \"cc5478a3-096b-4735-9c6e-8ff375821667\" (UID: \"cc5478a3-096b-4735-9c6e-8ff375821667\") " Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.851006 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dd72bd81-8786-47a1-81bd-9ae5d51cc8c2-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.851025 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dd72bd81-8786-47a1-81bd-9ae5d51cc8c2-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.851035 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nwh8r\" (UniqueName: \"kubernetes.io/projected/dd72bd81-8786-47a1-81bd-9ae5d51cc8c2-kube-api-access-nwh8r\") on node \"crc\" DevicePath \"\"" Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.851099 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cc5478a3-096b-4735-9c6e-8ff375821667-config" (OuterVolumeSpecName: "config") pod "cc5478a3-096b-4735-9c6e-8ff375821667" (UID: "cc5478a3-096b-4735-9c6e-8ff375821667"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.872803 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc5478a3-096b-4735-9c6e-8ff375821667-kube-api-access-rqgxz" (OuterVolumeSpecName: "kube-api-access-rqgxz") pod "cc5478a3-096b-4735-9c6e-8ff375821667" (UID: "cc5478a3-096b-4735-9c6e-8ff375821667"). InnerVolumeSpecName "kube-api-access-rqgxz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.952668 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rqgxz\" (UniqueName: \"kubernetes.io/projected/cc5478a3-096b-4735-9c6e-8ff375821667-kube-api-access-rqgxz\") on node \"crc\" DevicePath \"\"" Dec 09 15:47:45 crc kubenswrapper[4894]: I1209 15:47:45.952710 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cc5478a3-096b-4735-9c6e-8ff375821667-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:47:46 crc kubenswrapper[4894]: I1209 15:47:46.041975 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5f854695bc-cjt4j" event={"ID":"dd72bd81-8786-47a1-81bd-9ae5d51cc8c2","Type":"ContainerDied","Data":"92462fa45e0ea844173c746844b6b4990804ab18a03da9d8fe46b2123c6be857"} Dec 09 15:47:46 crc kubenswrapper[4894]: I1209 15:47:46.042051 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5f854695bc-cjt4j" Dec 09 15:47:46 crc kubenswrapper[4894]: I1209 15:47:46.045716 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4eebb217-538f-4f1a-a4f8-93371486803d","Type":"ContainerStarted","Data":"0054184f8422f9aecb07ba3936731bc91a0c623b6919b1ce18bbe55c86a01d1a"} Dec 09 15:47:46 crc kubenswrapper[4894]: I1209 15:47:46.051585 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"493425e2-b4c7-4a1b-af05-44b4f44bd2b3","Type":"ContainerStarted","Data":"e0cc90efd20a8df80e1313ad14be34aed239c406e39225fc092a39ec293bc945"} Dec 09 15:47:46 crc kubenswrapper[4894]: I1209 15:47:46.054787 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"44fdaebd-a96c-47f9-a421-a0984f78972e","Type":"ContainerStarted","Data":"89642012cdb0862001c73aa935177e0ea6e93168b497fc358d09b0a6cbda4b1a"} Dec 09 15:47:46 crc kubenswrapper[4894]: I1209 15:47:46.057622 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-84bb9d8bd9-4n4xr" event={"ID":"cc5478a3-096b-4735-9c6e-8ff375821667","Type":"ContainerDied","Data":"040f09540740fbe90c65dd001e6324418157ab60d9e87cad358c2271adad6eea"} Dec 09 15:47:46 crc kubenswrapper[4894]: I1209 15:47:46.057753 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-84bb9d8bd9-4n4xr" Dec 09 15:47:46 crc kubenswrapper[4894]: I1209 15:47:46.155454 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-cjt4j"] Dec 09 15:47:46 crc kubenswrapper[4894]: I1209 15:47:46.155493 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5f854695bc-cjt4j"] Dec 09 15:47:46 crc kubenswrapper[4894]: I1209 15:47:46.198074 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-4n4xr"] Dec 09 15:47:46 crc kubenswrapper[4894]: I1209 15:47:46.203472 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-84bb9d8bd9-4n4xr"] Dec 09 15:47:48 crc kubenswrapper[4894]: I1209 15:47:48.122546 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc5478a3-096b-4735-9c6e-8ff375821667" path="/var/lib/kubelet/pods/cc5478a3-096b-4735-9c6e-8ff375821667/volumes" Dec 09 15:47:48 crc kubenswrapper[4894]: I1209 15:47:48.123447 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd72bd81-8786-47a1-81bd-9ae5d51cc8c2" path="/var/lib/kubelet/pods/dd72bd81-8786-47a1-81bd-9ae5d51cc8c2/volumes" Dec 09 15:47:49 crc kubenswrapper[4894]: I1209 15:47:49.080995 4894 generic.go:334] "Generic (PLEG): container finished" podID="10d76c34-d0cc-493b-a57a-1923f8d72824" containerID="9485aa05471a006a1012a9a84b30857454c22e3295643dc269dd2ea3878d44cc" exitCode=0 Dec 09 15:47:49 crc kubenswrapper[4894]: I1209 15:47:49.081078 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"10d76c34-d0cc-493b-a57a-1923f8d72824","Type":"ContainerDied","Data":"9485aa05471a006a1012a9a84b30857454c22e3295643dc269dd2ea3878d44cc"} Dec 09 15:47:49 crc kubenswrapper[4894]: I1209 15:47:49.084090 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"5caa6181-4755-426f-97fa-50a47d8ea80a","Type":"ContainerDied","Data":"565ae530aed1596c22f18435e7871b4ab6e95db11efb296135dfe54ce89a5d75"} Dec 09 15:47:49 crc kubenswrapper[4894]: I1209 15:47:49.084112 4894 generic.go:334] "Generic (PLEG): container finished" podID="5caa6181-4755-426f-97fa-50a47d8ea80a" containerID="565ae530aed1596c22f18435e7871b4ab6e95db11efb296135dfe54ce89a5d75" exitCode=0 Dec 09 15:47:50 crc kubenswrapper[4894]: I1209 15:47:50.093803 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4eebb217-538f-4f1a-a4f8-93371486803d","Type":"ContainerStarted","Data":"85f9b5d998507b31a51e00147a68c908aa9e4b94f1931e7de79fcffca7d04346"} Dec 09 15:47:50 crc kubenswrapper[4894]: I1209 15:47:50.095488 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-z7vcg" event={"ID":"b0fedf20-9f17-4d49-bef9-c9476a5027cf","Type":"ContainerStarted","Data":"eb1c6337cffce12b3b4aa679e41af01fc063a411dea0d3b38dc1364725b2a87f"} Dec 09 15:47:50 crc kubenswrapper[4894]: I1209 15:47:50.098236 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"10d76c34-d0cc-493b-a57a-1923f8d72824","Type":"ContainerStarted","Data":"4a64f34bec454acdd68b0469ba60bfe17f4b4d901b3635eb69824da81fd5415a"} Dec 09 15:47:50 crc kubenswrapper[4894]: I1209 15:47:50.100611 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a6e0e621-ea8e-4256-8c04-74c54686708a","Type":"ContainerStarted","Data":"32c19ca4e792953a97f470c0a139b75e60ac0c64de54c2411a64a061a32ced7f"} Dec 09 15:47:50 crc kubenswrapper[4894]: I1209 15:47:50.102653 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"5caa6181-4755-426f-97fa-50a47d8ea80a","Type":"ContainerStarted","Data":"bec89fe05f2e1480a95701fc5ac8f41cc7d2f086c6ec790cede9b4db410f9da6"} Dec 09 15:47:50 crc kubenswrapper[4894]: I1209 15:47:50.104686 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vk5h6" event={"ID":"cf53d7df-bcc4-46b3-b874-dac48eca5bc4","Type":"ContainerStarted","Data":"ff524aa0b27eaabf039c5cb0be418640890a43850221c3e9ff06e14670529989"} Dec 09 15:47:50 crc kubenswrapper[4894]: I1209 15:47:50.105081 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-vk5h6" Dec 09 15:47:50 crc kubenswrapper[4894]: I1209 15:47:50.143416 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-vk5h6" podStartSLOduration=21.355291519 podStartE2EDuration="26.143399161s" podCreationTimestamp="2025-12-09 15:47:24 +0000 UTC" firstStartedPulling="2025-12-09 15:47:44.709131947 +0000 UTC m=+959.028342636" lastFinishedPulling="2025-12-09 15:47:49.497239609 +0000 UTC m=+963.816450278" observedRunningTime="2025-12-09 15:47:50.13593958 +0000 UTC m=+964.455150249" watchObservedRunningTime="2025-12-09 15:47:50.143399161 +0000 UTC m=+964.462609830" Dec 09 15:47:50 crc kubenswrapper[4894]: I1209 15:47:50.163787 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=8.324303576 podStartE2EDuration="35.16376944s" podCreationTimestamp="2025-12-09 15:47:15 +0000 UTC" firstStartedPulling="2025-12-09 15:47:17.436272537 +0000 UTC m=+931.755483206" lastFinishedPulling="2025-12-09 15:47:44.275738401 +0000 UTC m=+958.594949070" observedRunningTime="2025-12-09 15:47:50.15785661 +0000 UTC m=+964.477067279" watchObservedRunningTime="2025-12-09 15:47:50.16376944 +0000 UTC m=+964.482980109" Dec 09 15:47:50 crc kubenswrapper[4894]: I1209 15:47:50.215049 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=8.617983948 podStartE2EDuration="34.215026172s" podCreationTimestamp="2025-12-09 15:47:16 +0000 UTC" firstStartedPulling="2025-12-09 15:47:18.660029344 +0000 UTC m=+932.979240013" lastFinishedPulling="2025-12-09 15:47:44.257071568 +0000 UTC m=+958.576282237" observedRunningTime="2025-12-09 15:47:50.176538975 +0000 UTC m=+964.495749644" watchObservedRunningTime="2025-12-09 15:47:50.215026172 +0000 UTC m=+964.534236831" Dec 09 15:47:51 crc kubenswrapper[4894]: I1209 15:47:51.115439 4894 generic.go:334] "Generic (PLEG): container finished" podID="b0fedf20-9f17-4d49-bef9-c9476a5027cf" containerID="eb1c6337cffce12b3b4aa679e41af01fc063a411dea0d3b38dc1364725b2a87f" exitCode=0 Dec 09 15:47:51 crc kubenswrapper[4894]: I1209 15:47:51.117234 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-z7vcg" event={"ID":"b0fedf20-9f17-4d49-bef9-c9476a5027cf","Type":"ContainerDied","Data":"eb1c6337cffce12b3b4aa679e41af01fc063a411dea0d3b38dc1364725b2a87f"} Dec 09 15:47:53 crc kubenswrapper[4894]: I1209 15:47:53.132894 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"a6e0e621-ea8e-4256-8c04-74c54686708a","Type":"ContainerStarted","Data":"7177c58f0a1d1f67ad8755c7ba1b569ecbeaead8d04e2150160a8bd1bc1af0e6"} Dec 09 15:47:53 crc kubenswrapper[4894]: I1209 15:47:53.136092 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"48d401c8-b930-410e-a313-80399a664641","Type":"ContainerStarted","Data":"94f37c74641c85bdc8795734b07ff70ec2f8ac7e3c25edf2aac47eace970074d"} Dec 09 15:47:53 crc kubenswrapper[4894]: I1209 15:47:53.136232 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 09 15:47:53 crc kubenswrapper[4894]: I1209 15:47:53.138567 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"4eebb217-538f-4f1a-a4f8-93371486803d","Type":"ContainerStarted","Data":"4c86c970468c2a21bfa2a16ab1207e84beee3d138d1084249dd12a3e02bf1a5b"} Dec 09 15:47:53 crc kubenswrapper[4894]: I1209 15:47:53.142663 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-z7vcg" event={"ID":"b0fedf20-9f17-4d49-bef9-c9476a5027cf","Type":"ContainerStarted","Data":"d57b8aec5a1cc3ae910d8551bcc6f230603b283305001c5016f1d8e21e731357"} Dec 09 15:47:53 crc kubenswrapper[4894]: I1209 15:47:53.142710 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-z7vcg" event={"ID":"b0fedf20-9f17-4d49-bef9-c9476a5027cf","Type":"ContainerStarted","Data":"32e7047c59cb720aa5d8406aaed470e3229f94ac8f002177eae7b2bc378cc1f4"} Dec 09 15:47:53 crc kubenswrapper[4894]: I1209 15:47:53.143464 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:47:53 crc kubenswrapper[4894]: I1209 15:47:53.143505 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:47:53 crc kubenswrapper[4894]: I1209 15:47:53.157083 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=19.405675127 podStartE2EDuration="27.157051987s" podCreationTimestamp="2025-12-09 15:47:26 +0000 UTC" firstStartedPulling="2025-12-09 15:47:44.70961824 +0000 UTC m=+959.028828929" lastFinishedPulling="2025-12-09 15:47:52.46099512 +0000 UTC m=+966.780205789" observedRunningTime="2025-12-09 15:47:53.155946458 +0000 UTC m=+967.475157147" watchObservedRunningTime="2025-12-09 15:47:53.157051987 +0000 UTC m=+967.476262656" Dec 09 15:47:53 crc kubenswrapper[4894]: I1209 15:47:53.207623 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-z7vcg" podStartSLOduration=24.000194053 podStartE2EDuration="29.20760468s" podCreationTimestamp="2025-12-09 15:47:24 +0000 UTC" firstStartedPulling="2025-12-09 15:47:44.264316333 +0000 UTC m=+958.583527002" lastFinishedPulling="2025-12-09 15:47:49.47172696 +0000 UTC m=+963.790937629" observedRunningTime="2025-12-09 15:47:53.205816552 +0000 UTC m=+967.525027231" watchObservedRunningTime="2025-12-09 15:47:53.20760468 +0000 UTC m=+967.526815349" Dec 09 15:47:53 crc kubenswrapper[4894]: I1209 15:47:53.208191 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=19.906665026 podStartE2EDuration="27.208183057s" podCreationTimestamp="2025-12-09 15:47:26 +0000 UTC" firstStartedPulling="2025-12-09 15:47:45.164931606 +0000 UTC m=+959.484142275" lastFinishedPulling="2025-12-09 15:47:52.466449637 +0000 UTC m=+966.785660306" observedRunningTime="2025-12-09 15:47:53.183199473 +0000 UTC m=+967.502410152" watchObservedRunningTime="2025-12-09 15:47:53.208183057 +0000 UTC m=+967.527393746" Dec 09 15:47:53 crc kubenswrapper[4894]: I1209 15:47:53.224731 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=25.100423187 podStartE2EDuration="33.224714643s" podCreationTimestamp="2025-12-09 15:47:20 +0000 UTC" firstStartedPulling="2025-12-09 15:47:44.264664172 +0000 UTC m=+958.583874841" lastFinishedPulling="2025-12-09 15:47:52.388955628 +0000 UTC m=+966.708166297" observedRunningTime="2025-12-09 15:47:53.220037236 +0000 UTC m=+967.539247905" watchObservedRunningTime="2025-12-09 15:47:53.224714643 +0000 UTC m=+967.543925312" Dec 09 15:47:53 crc kubenswrapper[4894]: I1209 15:47:53.459763 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:54.668914 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:54.718736 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:54.910961 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:54.956593 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.156024 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.156062 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.196750 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.198202 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.486604 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-744ffd65bc-2hr9g"] Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.533483 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7878659675-bwbdn"] Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.535199 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7878659675-bwbdn" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.542484 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.545558 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7878659675-bwbdn"] Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.599796 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-459zv"] Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.601632 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-459zv" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.604393 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.631561 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/963becb8-0d9d-48a5-b11f-29f1dd7ebb7d-ovs-rundir\") pod \"ovn-controller-metrics-459zv\" (UID: \"963becb8-0d9d-48a5-b11f-29f1dd7ebb7d\") " pod="openstack/ovn-controller-metrics-459zv" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.631602 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9zs7j\" (UniqueName: \"kubernetes.io/projected/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-kube-api-access-9zs7j\") pod \"dnsmasq-dns-7878659675-bwbdn\" (UID: \"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad\") " pod="openstack/dnsmasq-dns-7878659675-bwbdn" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.631626 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/963becb8-0d9d-48a5-b11f-29f1dd7ebb7d-config\") pod \"ovn-controller-metrics-459zv\" (UID: \"963becb8-0d9d-48a5-b11f-29f1dd7ebb7d\") " pod="openstack/ovn-controller-metrics-459zv" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.631670 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-config\") pod \"dnsmasq-dns-7878659675-bwbdn\" (UID: \"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad\") " pod="openstack/dnsmasq-dns-7878659675-bwbdn" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.631688 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/963becb8-0d9d-48a5-b11f-29f1dd7ebb7d-ovn-rundir\") pod \"ovn-controller-metrics-459zv\" (UID: \"963becb8-0d9d-48a5-b11f-29f1dd7ebb7d\") " pod="openstack/ovn-controller-metrics-459zv" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.631720 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/963becb8-0d9d-48a5-b11f-29f1dd7ebb7d-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-459zv\" (UID: \"963becb8-0d9d-48a5-b11f-29f1dd7ebb7d\") " pod="openstack/ovn-controller-metrics-459zv" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.631744 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-ovsdbserver-nb\") pod \"dnsmasq-dns-7878659675-bwbdn\" (UID: \"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad\") " pod="openstack/dnsmasq-dns-7878659675-bwbdn" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.631762 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-dns-svc\") pod \"dnsmasq-dns-7878659675-bwbdn\" (UID: \"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad\") " pod="openstack/dnsmasq-dns-7878659675-bwbdn" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.631785 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqm6c\" (UniqueName: \"kubernetes.io/projected/963becb8-0d9d-48a5-b11f-29f1dd7ebb7d-kube-api-access-gqm6c\") pod \"ovn-controller-metrics-459zv\" (UID: \"963becb8-0d9d-48a5-b11f-29f1dd7ebb7d\") " pod="openstack/ovn-controller-metrics-459zv" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.631809 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/963becb8-0d9d-48a5-b11f-29f1dd7ebb7d-combined-ca-bundle\") pod \"ovn-controller-metrics-459zv\" (UID: \"963becb8-0d9d-48a5-b11f-29f1dd7ebb7d\") " pod="openstack/ovn-controller-metrics-459zv" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.655220 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-459zv"] Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.694543 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-sk7xs"] Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.732027 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-586b989cdc-946tp"] Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.732999 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/963becb8-0d9d-48a5-b11f-29f1dd7ebb7d-ovs-rundir\") pod \"ovn-controller-metrics-459zv\" (UID: \"963becb8-0d9d-48a5-b11f-29f1dd7ebb7d\") " pod="openstack/ovn-controller-metrics-459zv" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.733049 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9zs7j\" (UniqueName: \"kubernetes.io/projected/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-kube-api-access-9zs7j\") pod \"dnsmasq-dns-7878659675-bwbdn\" (UID: \"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad\") " pod="openstack/dnsmasq-dns-7878659675-bwbdn" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.733078 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/963becb8-0d9d-48a5-b11f-29f1dd7ebb7d-config\") pod \"ovn-controller-metrics-459zv\" (UID: \"963becb8-0d9d-48a5-b11f-29f1dd7ebb7d\") " pod="openstack/ovn-controller-metrics-459zv" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.733104 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-config\") pod \"dnsmasq-dns-7878659675-bwbdn\" (UID: \"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad\") " pod="openstack/dnsmasq-dns-7878659675-bwbdn" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.741071 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/963becb8-0d9d-48a5-b11f-29f1dd7ebb7d-ovn-rundir\") pod \"ovn-controller-metrics-459zv\" (UID: \"963becb8-0d9d-48a5-b11f-29f1dd7ebb7d\") " pod="openstack/ovn-controller-metrics-459zv" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.741156 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/963becb8-0d9d-48a5-b11f-29f1dd7ebb7d-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-459zv\" (UID: \"963becb8-0d9d-48a5-b11f-29f1dd7ebb7d\") " pod="openstack/ovn-controller-metrics-459zv" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.741204 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-ovsdbserver-nb\") pod \"dnsmasq-dns-7878659675-bwbdn\" (UID: \"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad\") " pod="openstack/dnsmasq-dns-7878659675-bwbdn" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.741336 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-dns-svc\") pod \"dnsmasq-dns-7878659675-bwbdn\" (UID: \"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad\") " pod="openstack/dnsmasq-dns-7878659675-bwbdn" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.741366 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqm6c\" (UniqueName: \"kubernetes.io/projected/963becb8-0d9d-48a5-b11f-29f1dd7ebb7d-kube-api-access-gqm6c\") pod \"ovn-controller-metrics-459zv\" (UID: \"963becb8-0d9d-48a5-b11f-29f1dd7ebb7d\") " pod="openstack/ovn-controller-metrics-459zv" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.741396 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/963becb8-0d9d-48a5-b11f-29f1dd7ebb7d-combined-ca-bundle\") pod \"ovn-controller-metrics-459zv\" (UID: \"963becb8-0d9d-48a5-b11f-29f1dd7ebb7d\") " pod="openstack/ovn-controller-metrics-459zv" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.747273 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/963becb8-0d9d-48a5-b11f-29f1dd7ebb7d-ovn-rundir\") pod \"ovn-controller-metrics-459zv\" (UID: \"963becb8-0d9d-48a5-b11f-29f1dd7ebb7d\") " pod="openstack/ovn-controller-metrics-459zv" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.747555 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/963becb8-0d9d-48a5-b11f-29f1dd7ebb7d-config\") pod \"ovn-controller-metrics-459zv\" (UID: \"963becb8-0d9d-48a5-b11f-29f1dd7ebb7d\") " pod="openstack/ovn-controller-metrics-459zv" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.747717 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/963becb8-0d9d-48a5-b11f-29f1dd7ebb7d-ovs-rundir\") pod \"ovn-controller-metrics-459zv\" (UID: \"963becb8-0d9d-48a5-b11f-29f1dd7ebb7d\") " pod="openstack/ovn-controller-metrics-459zv" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.748138 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-ovsdbserver-nb\") pod \"dnsmasq-dns-7878659675-bwbdn\" (UID: \"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad\") " pod="openstack/dnsmasq-dns-7878659675-bwbdn" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.748350 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-config\") pod \"dnsmasq-dns-7878659675-bwbdn\" (UID: \"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad\") " pod="openstack/dnsmasq-dns-7878659675-bwbdn" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.748496 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586b989cdc-946tp" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.748708 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-dns-svc\") pod \"dnsmasq-dns-7878659675-bwbdn\" (UID: \"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad\") " pod="openstack/dnsmasq-dns-7878659675-bwbdn" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.753730 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.770579 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/963becb8-0d9d-48a5-b11f-29f1dd7ebb7d-combined-ca-bundle\") pod \"ovn-controller-metrics-459zv\" (UID: \"963becb8-0d9d-48a5-b11f-29f1dd7ebb7d\") " pod="openstack/ovn-controller-metrics-459zv" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.787441 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-586b989cdc-946tp"] Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.796953 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.799327 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.810599 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9zs7j\" (UniqueName: \"kubernetes.io/projected/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-kube-api-access-9zs7j\") pod \"dnsmasq-dns-7878659675-bwbdn\" (UID: \"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad\") " pod="openstack/dnsmasq-dns-7878659675-bwbdn" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.815901 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqm6c\" (UniqueName: \"kubernetes.io/projected/963becb8-0d9d-48a5-b11f-29f1dd7ebb7d-kube-api-access-gqm6c\") pod \"ovn-controller-metrics-459zv\" (UID: \"963becb8-0d9d-48a5-b11f-29f1dd7ebb7d\") " pod="openstack/ovn-controller-metrics-459zv" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.817969 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/963becb8-0d9d-48a5-b11f-29f1dd7ebb7d-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-459zv\" (UID: \"963becb8-0d9d-48a5-b11f-29f1dd7ebb7d\") " pod="openstack/ovn-controller-metrics-459zv" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.840457 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.840730 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.840830 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.840928 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-fvlvs" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.843213 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-ovsdbserver-nb\") pod \"dnsmasq-dns-586b989cdc-946tp\" (UID: \"640684c9-0085-424c-a19d-3510625cef05\") " pod="openstack/dnsmasq-dns-586b989cdc-946tp" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.843274 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4e050ff-6a42-40ac-9298-f87ce2f7743a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c4e050ff-6a42-40ac-9298-f87ce2f7743a\") " pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.843306 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nncbx\" (UniqueName: \"kubernetes.io/projected/640684c9-0085-424c-a19d-3510625cef05-kube-api-access-nncbx\") pod \"dnsmasq-dns-586b989cdc-946tp\" (UID: \"640684c9-0085-424c-a19d-3510625cef05\") " pod="openstack/dnsmasq-dns-586b989cdc-946tp" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.843345 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-config\") pod \"dnsmasq-dns-586b989cdc-946tp\" (UID: \"640684c9-0085-424c-a19d-3510625cef05\") " pod="openstack/dnsmasq-dns-586b989cdc-946tp" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.843396 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-dns-svc\") pod \"dnsmasq-dns-586b989cdc-946tp\" (UID: \"640684c9-0085-424c-a19d-3510625cef05\") " pod="openstack/dnsmasq-dns-586b989cdc-946tp" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.843447 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-ovsdbserver-sb\") pod \"dnsmasq-dns-586b989cdc-946tp\" (UID: \"640684c9-0085-424c-a19d-3510625cef05\") " pod="openstack/dnsmasq-dns-586b989cdc-946tp" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.843533 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4e050ff-6a42-40ac-9298-f87ce2f7743a-config\") pod \"ovn-northd-0\" (UID: \"c4e050ff-6a42-40ac-9298-f87ce2f7743a\") " pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.843695 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c4e050ff-6a42-40ac-9298-f87ce2f7743a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c4e050ff-6a42-40ac-9298-f87ce2f7743a\") " pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.843778 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4e050ff-6a42-40ac-9298-f87ce2f7743a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c4e050ff-6a42-40ac-9298-f87ce2f7743a\") " pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.843863 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4e050ff-6a42-40ac-9298-f87ce2f7743a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c4e050ff-6a42-40ac-9298-f87ce2f7743a\") " pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.843933 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c4e050ff-6a42-40ac-9298-f87ce2f7743a-scripts\") pod \"ovn-northd-0\" (UID: \"c4e050ff-6a42-40ac-9298-f87ce2f7743a\") " pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.843980 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kr9cf\" (UniqueName: \"kubernetes.io/projected/c4e050ff-6a42-40ac-9298-f87ce2f7743a-kube-api-access-kr9cf\") pod \"ovn-northd-0\" (UID: \"c4e050ff-6a42-40ac-9298-f87ce2f7743a\") " pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.855417 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.861256 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7878659675-bwbdn" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.923555 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-459zv" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.948894 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-ovsdbserver-nb\") pod \"dnsmasq-dns-586b989cdc-946tp\" (UID: \"640684c9-0085-424c-a19d-3510625cef05\") " pod="openstack/dnsmasq-dns-586b989cdc-946tp" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.948962 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4e050ff-6a42-40ac-9298-f87ce2f7743a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c4e050ff-6a42-40ac-9298-f87ce2f7743a\") " pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.949013 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nncbx\" (UniqueName: \"kubernetes.io/projected/640684c9-0085-424c-a19d-3510625cef05-kube-api-access-nncbx\") pod \"dnsmasq-dns-586b989cdc-946tp\" (UID: \"640684c9-0085-424c-a19d-3510625cef05\") " pod="openstack/dnsmasq-dns-586b989cdc-946tp" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.949056 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-config\") pod \"dnsmasq-dns-586b989cdc-946tp\" (UID: \"640684c9-0085-424c-a19d-3510625cef05\") " pod="openstack/dnsmasq-dns-586b989cdc-946tp" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.949106 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-dns-svc\") pod \"dnsmasq-dns-586b989cdc-946tp\" (UID: \"640684c9-0085-424c-a19d-3510625cef05\") " pod="openstack/dnsmasq-dns-586b989cdc-946tp" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.949141 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-ovsdbserver-sb\") pod \"dnsmasq-dns-586b989cdc-946tp\" (UID: \"640684c9-0085-424c-a19d-3510625cef05\") " pod="openstack/dnsmasq-dns-586b989cdc-946tp" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.949166 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4e050ff-6a42-40ac-9298-f87ce2f7743a-config\") pod \"ovn-northd-0\" (UID: \"c4e050ff-6a42-40ac-9298-f87ce2f7743a\") " pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.949204 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c4e050ff-6a42-40ac-9298-f87ce2f7743a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c4e050ff-6a42-40ac-9298-f87ce2f7743a\") " pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.949246 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4e050ff-6a42-40ac-9298-f87ce2f7743a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c4e050ff-6a42-40ac-9298-f87ce2f7743a\") " pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.949286 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4e050ff-6a42-40ac-9298-f87ce2f7743a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c4e050ff-6a42-40ac-9298-f87ce2f7743a\") " pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.949327 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c4e050ff-6a42-40ac-9298-f87ce2f7743a-scripts\") pod \"ovn-northd-0\" (UID: \"c4e050ff-6a42-40ac-9298-f87ce2f7743a\") " pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.949358 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kr9cf\" (UniqueName: \"kubernetes.io/projected/c4e050ff-6a42-40ac-9298-f87ce2f7743a-kube-api-access-kr9cf\") pod \"ovn-northd-0\" (UID: \"c4e050ff-6a42-40ac-9298-f87ce2f7743a\") " pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.950318 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/c4e050ff-6a42-40ac-9298-f87ce2f7743a-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"c4e050ff-6a42-40ac-9298-f87ce2f7743a\") " pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.950548 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-dns-svc\") pod \"dnsmasq-dns-586b989cdc-946tp\" (UID: \"640684c9-0085-424c-a19d-3510625cef05\") " pod="openstack/dnsmasq-dns-586b989cdc-946tp" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.950973 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c4e050ff-6a42-40ac-9298-f87ce2f7743a-config\") pod \"ovn-northd-0\" (UID: \"c4e050ff-6a42-40ac-9298-f87ce2f7743a\") " pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.952453 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c4e050ff-6a42-40ac-9298-f87ce2f7743a-scripts\") pod \"ovn-northd-0\" (UID: \"c4e050ff-6a42-40ac-9298-f87ce2f7743a\") " pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.953029 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-config\") pod \"dnsmasq-dns-586b989cdc-946tp\" (UID: \"640684c9-0085-424c-a19d-3510625cef05\") " pod="openstack/dnsmasq-dns-586b989cdc-946tp" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.953164 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-ovsdbserver-sb\") pod \"dnsmasq-dns-586b989cdc-946tp\" (UID: \"640684c9-0085-424c-a19d-3510625cef05\") " pod="openstack/dnsmasq-dns-586b989cdc-946tp" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.953747 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-ovsdbserver-nb\") pod \"dnsmasq-dns-586b989cdc-946tp\" (UID: \"640684c9-0085-424c-a19d-3510625cef05\") " pod="openstack/dnsmasq-dns-586b989cdc-946tp" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.954484 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c4e050ff-6a42-40ac-9298-f87ce2f7743a-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"c4e050ff-6a42-40ac-9298-f87ce2f7743a\") " pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.954725 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4e050ff-6a42-40ac-9298-f87ce2f7743a-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"c4e050ff-6a42-40ac-9298-f87ce2f7743a\") " pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.964404 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/c4e050ff-6a42-40ac-9298-f87ce2f7743a-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"c4e050ff-6a42-40ac-9298-f87ce2f7743a\") " pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.968920 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kr9cf\" (UniqueName: \"kubernetes.io/projected/c4e050ff-6a42-40ac-9298-f87ce2f7743a-kube-api-access-kr9cf\") pod \"ovn-northd-0\" (UID: \"c4e050ff-6a42-40ac-9298-f87ce2f7743a\") " pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.975538 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nncbx\" (UniqueName: \"kubernetes.io/projected/640684c9-0085-424c-a19d-3510625cef05-kube-api-access-nncbx\") pod \"dnsmasq-dns-586b989cdc-946tp\" (UID: \"640684c9-0085-424c-a19d-3510625cef05\") " pod="openstack/dnsmasq-dns-586b989cdc-946tp" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.991784 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:55.999811 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744ffd65bc-2hr9g" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:56.050460 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f33e7674-a9a6-4116-807c-9d35ad260b6b-dns-svc\") pod \"f33e7674-a9a6-4116-807c-9d35ad260b6b\" (UID: \"f33e7674-a9a6-4116-807c-9d35ad260b6b\") " Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:56.050499 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f33e7674-a9a6-4116-807c-9d35ad260b6b-config\") pod \"f33e7674-a9a6-4116-807c-9d35ad260b6b\" (UID: \"f33e7674-a9a6-4116-807c-9d35ad260b6b\") " Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:56.050561 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7qc74\" (UniqueName: \"kubernetes.io/projected/f33e7674-a9a6-4116-807c-9d35ad260b6b-kube-api-access-7qc74\") pod \"f33e7674-a9a6-4116-807c-9d35ad260b6b\" (UID: \"f33e7674-a9a6-4116-807c-9d35ad260b6b\") " Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:56.050972 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f33e7674-a9a6-4116-807c-9d35ad260b6b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f33e7674-a9a6-4116-807c-9d35ad260b6b" (UID: "f33e7674-a9a6-4116-807c-9d35ad260b6b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:56.051966 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f33e7674-a9a6-4116-807c-9d35ad260b6b-config" (OuterVolumeSpecName: "config") pod "f33e7674-a9a6-4116-807c-9d35ad260b6b" (UID: "f33e7674-a9a6-4116-807c-9d35ad260b6b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:56.058290 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f33e7674-a9a6-4116-807c-9d35ad260b6b-kube-api-access-7qc74" (OuterVolumeSpecName: "kube-api-access-7qc74") pod "f33e7674-a9a6-4116-807c-9d35ad260b6b" (UID: "f33e7674-a9a6-4116-807c-9d35ad260b6b"). InnerVolumeSpecName "kube-api-access-7qc74". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:56.153262 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f33e7674-a9a6-4116-807c-9d35ad260b6b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:56.153295 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f33e7674-a9a6-4116-807c-9d35ad260b6b-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:56.153307 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7qc74\" (UniqueName: \"kubernetes.io/projected/f33e7674-a9a6-4116-807c-9d35ad260b6b-kube-api-access-7qc74\") on node \"crc\" DevicePath \"\"" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:56.167588 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-744ffd65bc-2hr9g" event={"ID":"f33e7674-a9a6-4116-807c-9d35ad260b6b","Type":"ContainerDied","Data":"c048271f44b28d2dcad20b623a726c775219abd8a6f428b2c6593cd9e7a45301"} Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:56.167725 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-744ffd65bc-2hr9g" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:56.211293 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-744ffd65bc-2hr9g"] Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:56.212946 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586b989cdc-946tp" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:56.220576 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-744ffd65bc-2hr9g"] Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:56.759469 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:56.759517 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:58.049156 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:58.049490 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:58.118771 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f33e7674-a9a6-4116-807c-9d35ad260b6b" path="/var/lib/kubelet/pods/f33e7674-a9a6-4116-807c-9d35ad260b6b/volumes" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:58.845344 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-95f5f6995-sk7xs" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:58.904272 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-459zv"] Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:58.911338 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:58.934480 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-586b989cdc-946tp"] Dec 09 15:47:58 crc kubenswrapper[4894]: W1209 15:47:58.953802 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7dfad8b3_cbaa_4a80_ae85_799b1fa7f7ad.slice/crio-a0fdadc383b2f2b90bcded1445ddccf8bfd18a72e2ebe60207251ee012fc1455 WatchSource:0}: Error finding container a0fdadc383b2f2b90bcded1445ddccf8bfd18a72e2ebe60207251ee012fc1455: Status 404 returned error can't find the container with id a0fdadc383b2f2b90bcded1445ddccf8bfd18a72e2ebe60207251ee012fc1455 Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:58.955010 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7878659675-bwbdn"] Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:58.996826 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gbqkj\" (UniqueName: \"kubernetes.io/projected/73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb-kube-api-access-gbqkj\") pod \"73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb\" (UID: \"73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb\") " Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:58.997235 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb-dns-svc\") pod \"73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb\" (UID: \"73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb\") " Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:58.997337 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb-config\") pod \"73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb\" (UID: \"73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb\") " Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:58.998184 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb-config" (OuterVolumeSpecName: "config") pod "73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb" (UID: "73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:47:58 crc kubenswrapper[4894]: I1209 15:47:58.998234 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb" (UID: "73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:47:59 crc kubenswrapper[4894]: I1209 15:47:59.000237 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb-kube-api-access-gbqkj" (OuterVolumeSpecName: "kube-api-access-gbqkj") pod "73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb" (UID: "73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb"). InnerVolumeSpecName "kube-api-access-gbqkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:47:59 crc kubenswrapper[4894]: I1209 15:47:59.099347 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gbqkj\" (UniqueName: \"kubernetes.io/projected/73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb-kube-api-access-gbqkj\") on node \"crc\" DevicePath \"\"" Dec 09 15:47:59 crc kubenswrapper[4894]: I1209 15:47:59.099386 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 15:47:59 crc kubenswrapper[4894]: I1209 15:47:59.099398 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:47:59 crc kubenswrapper[4894]: I1209 15:47:59.192502 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-459zv" event={"ID":"963becb8-0d9d-48a5-b11f-29f1dd7ebb7d","Type":"ContainerStarted","Data":"f4178aaa26d5b28eead2c43a89d94ddca6e3f7a13328bf17bda0a1e90fb6388d"} Dec 09 15:47:59 crc kubenswrapper[4894]: I1209 15:47:59.193289 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c4e050ff-6a42-40ac-9298-f87ce2f7743a","Type":"ContainerStarted","Data":"f06f439af175c95ef14221c0c2da0f16b5a0554b1c51a3b62d6976479e543d23"} Dec 09 15:47:59 crc kubenswrapper[4894]: I1209 15:47:59.194139 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-95f5f6995-sk7xs" event={"ID":"73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb","Type":"ContainerDied","Data":"48c29f56bec5cdb814701b023a681e0eecf3e0dd1eb84bdb4e49063ee2c4e43c"} Dec 09 15:47:59 crc kubenswrapper[4894]: I1209 15:47:59.194181 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-95f5f6995-sk7xs" Dec 09 15:47:59 crc kubenswrapper[4894]: I1209 15:47:59.194938 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7878659675-bwbdn" event={"ID":"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad","Type":"ContainerStarted","Data":"a0fdadc383b2f2b90bcded1445ddccf8bfd18a72e2ebe60207251ee012fc1455"} Dec 09 15:47:59 crc kubenswrapper[4894]: I1209 15:47:59.195724 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586b989cdc-946tp" event={"ID":"640684c9-0085-424c-a19d-3510625cef05","Type":"ContainerStarted","Data":"6213b765214edb517bb6d70e8fbc638808980173618f278fa8f9837c25410b75"} Dec 09 15:47:59 crc kubenswrapper[4894]: I1209 15:47:59.244990 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-sk7xs"] Dec 09 15:47:59 crc kubenswrapper[4894]: I1209 15:47:59.251269 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-95f5f6995-sk7xs"] Dec 09 15:48:00 crc kubenswrapper[4894]: I1209 15:48:00.123693 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb" path="/var/lib/kubelet/pods/73245da2-1e1c-4a4d-ac6e-fe1ac0e56bcb/volumes" Dec 09 15:48:00 crc kubenswrapper[4894]: I1209 15:48:00.205117 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-459zv" event={"ID":"963becb8-0d9d-48a5-b11f-29f1dd7ebb7d","Type":"ContainerStarted","Data":"e71e0657f518f12a3a7605d0b1fd11b2afaa535ecd0f725f1e3164632adaf24d"} Dec 09 15:48:00 crc kubenswrapper[4894]: I1209 15:48:00.437673 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 09 15:48:02 crc kubenswrapper[4894]: I1209 15:48:02.641383 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Dec 09 15:48:02 crc kubenswrapper[4894]: I1209 15:48:02.668796 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-459zv" podStartSLOduration=7.6687710639999995 podStartE2EDuration="7.668771064s" podCreationTimestamp="2025-12-09 15:47:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:48:01.23040125 +0000 UTC m=+975.549611939" watchObservedRunningTime="2025-12-09 15:48:02.668771064 +0000 UTC m=+976.987981743" Dec 09 15:48:02 crc kubenswrapper[4894]: I1209 15:48:02.756144 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Dec 09 15:48:03 crc kubenswrapper[4894]: I1209 15:48:03.232549 4894 generic.go:334] "Generic (PLEG): container finished" podID="640684c9-0085-424c-a19d-3510625cef05" containerID="c17f44953076f2c6ef3d3fa75facd18c28928a9d052516faa5cb846118da9409" exitCode=0 Dec 09 15:48:03 crc kubenswrapper[4894]: I1209 15:48:03.232663 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586b989cdc-946tp" event={"ID":"640684c9-0085-424c-a19d-3510625cef05","Type":"ContainerDied","Data":"c17f44953076f2c6ef3d3fa75facd18c28928a9d052516faa5cb846118da9409"} Dec 09 15:48:03 crc kubenswrapper[4894]: I1209 15:48:03.235965 4894 generic.go:334] "Generic (PLEG): container finished" podID="7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad" containerID="bc6f9ffb28b4541dc62860082365a7be6cc5d642dc81599c8989a128bc35624c" exitCode=0 Dec 09 15:48:03 crc kubenswrapper[4894]: I1209 15:48:03.236699 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7878659675-bwbdn" event={"ID":"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad","Type":"ContainerDied","Data":"bc6f9ffb28b4541dc62860082365a7be6cc5d642dc81599c8989a128bc35624c"} Dec 09 15:48:03 crc kubenswrapper[4894]: I1209 15:48:03.677876 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-11cf-account-create-update-ggv4k"] Dec 09 15:48:03 crc kubenswrapper[4894]: I1209 15:48:03.701687 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-11cf-account-create-update-ggv4k"] Dec 09 15:48:03 crc kubenswrapper[4894]: I1209 15:48:03.701820 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-11cf-account-create-update-ggv4k" Dec 09 15:48:03 crc kubenswrapper[4894]: I1209 15:48:03.707584 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Dec 09 15:48:03 crc kubenswrapper[4894]: I1209 15:48:03.724549 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-hp2h5"] Dec 09 15:48:03 crc kubenswrapper[4894]: I1209 15:48:03.726218 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hp2h5" Dec 09 15:48:03 crc kubenswrapper[4894]: I1209 15:48:03.737836 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-hp2h5"] Dec 09 15:48:03 crc kubenswrapper[4894]: I1209 15:48:03.790255 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d962c43-b57b-4db6-bda2-c6b48bb7335a-operator-scripts\") pod \"glance-11cf-account-create-update-ggv4k\" (UID: \"9d962c43-b57b-4db6-bda2-c6b48bb7335a\") " pod="openstack/glance-11cf-account-create-update-ggv4k" Dec 09 15:48:03 crc kubenswrapper[4894]: I1209 15:48:03.790339 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hx5kl\" (UniqueName: \"kubernetes.io/projected/9d962c43-b57b-4db6-bda2-c6b48bb7335a-kube-api-access-hx5kl\") pod \"glance-11cf-account-create-update-ggv4k\" (UID: \"9d962c43-b57b-4db6-bda2-c6b48bb7335a\") " pod="openstack/glance-11cf-account-create-update-ggv4k" Dec 09 15:48:03 crc kubenswrapper[4894]: I1209 15:48:03.892475 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d962c43-b57b-4db6-bda2-c6b48bb7335a-operator-scripts\") pod \"glance-11cf-account-create-update-ggv4k\" (UID: \"9d962c43-b57b-4db6-bda2-c6b48bb7335a\") " pod="openstack/glance-11cf-account-create-update-ggv4k" Dec 09 15:48:03 crc kubenswrapper[4894]: I1209 15:48:03.893005 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hx5kl\" (UniqueName: \"kubernetes.io/projected/9d962c43-b57b-4db6-bda2-c6b48bb7335a-kube-api-access-hx5kl\") pod \"glance-11cf-account-create-update-ggv4k\" (UID: \"9d962c43-b57b-4db6-bda2-c6b48bb7335a\") " pod="openstack/glance-11cf-account-create-update-ggv4k" Dec 09 15:48:03 crc kubenswrapper[4894]: I1209 15:48:03.893173 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b-operator-scripts\") pod \"glance-db-create-hp2h5\" (UID: \"f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b\") " pod="openstack/glance-db-create-hp2h5" Dec 09 15:48:03 crc kubenswrapper[4894]: I1209 15:48:03.893283 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x8sfz\" (UniqueName: \"kubernetes.io/projected/f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b-kube-api-access-x8sfz\") pod \"glance-db-create-hp2h5\" (UID: \"f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b\") " pod="openstack/glance-db-create-hp2h5" Dec 09 15:48:03 crc kubenswrapper[4894]: I1209 15:48:03.893474 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d962c43-b57b-4db6-bda2-c6b48bb7335a-operator-scripts\") pod \"glance-11cf-account-create-update-ggv4k\" (UID: \"9d962c43-b57b-4db6-bda2-c6b48bb7335a\") " pod="openstack/glance-11cf-account-create-update-ggv4k" Dec 09 15:48:03 crc kubenswrapper[4894]: I1209 15:48:03.909414 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hx5kl\" (UniqueName: \"kubernetes.io/projected/9d962c43-b57b-4db6-bda2-c6b48bb7335a-kube-api-access-hx5kl\") pod \"glance-11cf-account-create-update-ggv4k\" (UID: \"9d962c43-b57b-4db6-bda2-c6b48bb7335a\") " pod="openstack/glance-11cf-account-create-update-ggv4k" Dec 09 15:48:03 crc kubenswrapper[4894]: I1209 15:48:03.995028 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x8sfz\" (UniqueName: \"kubernetes.io/projected/f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b-kube-api-access-x8sfz\") pod \"glance-db-create-hp2h5\" (UID: \"f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b\") " pod="openstack/glance-db-create-hp2h5" Dec 09 15:48:03 crc kubenswrapper[4894]: I1209 15:48:03.995186 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b-operator-scripts\") pod \"glance-db-create-hp2h5\" (UID: \"f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b\") " pod="openstack/glance-db-create-hp2h5" Dec 09 15:48:03 crc kubenswrapper[4894]: I1209 15:48:03.996025 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b-operator-scripts\") pod \"glance-db-create-hp2h5\" (UID: \"f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b\") " pod="openstack/glance-db-create-hp2h5" Dec 09 15:48:04 crc kubenswrapper[4894]: I1209 15:48:04.015412 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x8sfz\" (UniqueName: \"kubernetes.io/projected/f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b-kube-api-access-x8sfz\") pod \"glance-db-create-hp2h5\" (UID: \"f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b\") " pod="openstack/glance-db-create-hp2h5" Dec 09 15:48:04 crc kubenswrapper[4894]: I1209 15:48:04.068963 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-11cf-account-create-update-ggv4k" Dec 09 15:48:04 crc kubenswrapper[4894]: I1209 15:48:04.074314 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hp2h5" Dec 09 15:48:04 crc kubenswrapper[4894]: I1209 15:48:04.249042 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7878659675-bwbdn" event={"ID":"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad","Type":"ContainerStarted","Data":"aa3a6add204fac867d21e0c55ff60c6bf702cb003d34594c10dc1afc37493194"} Dec 09 15:48:04 crc kubenswrapper[4894]: I1209 15:48:04.249513 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7878659675-bwbdn" Dec 09 15:48:04 crc kubenswrapper[4894]: I1209 15:48:04.250933 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586b989cdc-946tp" event={"ID":"640684c9-0085-424c-a19d-3510625cef05","Type":"ContainerStarted","Data":"073c895aeb5d86d84883745aa5ee2682a9db1f2d05a9348e1633116b21b984d0"} Dec 09 15:48:04 crc kubenswrapper[4894]: I1209 15:48:04.251277 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-586b989cdc-946tp" Dec 09 15:48:04 crc kubenswrapper[4894]: I1209 15:48:04.255138 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c4e050ff-6a42-40ac-9298-f87ce2f7743a","Type":"ContainerStarted","Data":"e24549e05621220bbecafcc131396a2faede30e2fcfc43e799a418c720b74c29"} Dec 09 15:48:04 crc kubenswrapper[4894]: I1209 15:48:04.255221 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"c4e050ff-6a42-40ac-9298-f87ce2f7743a","Type":"ContainerStarted","Data":"7c26cd27f10e967a1419ad6daa50b56a4c90b8dcd6d746e5da17f2f06cdee6f2"} Dec 09 15:48:04 crc kubenswrapper[4894]: I1209 15:48:04.255310 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Dec 09 15:48:04 crc kubenswrapper[4894]: I1209 15:48:04.269933 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7878659675-bwbdn" podStartSLOduration=5.55963785 podStartE2EDuration="9.269917161s" podCreationTimestamp="2025-12-09 15:47:55 +0000 UTC" firstStartedPulling="2025-12-09 15:47:58.957415825 +0000 UTC m=+973.276626494" lastFinishedPulling="2025-12-09 15:48:02.667695126 +0000 UTC m=+976.986905805" observedRunningTime="2025-12-09 15:48:04.266345638 +0000 UTC m=+978.585556317" watchObservedRunningTime="2025-12-09 15:48:04.269917161 +0000 UTC m=+978.589127830" Dec 09 15:48:04 crc kubenswrapper[4894]: I1209 15:48:04.292012 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-586b989cdc-946tp" podStartSLOduration=5.582632331 podStartE2EDuration="9.291992808s" podCreationTimestamp="2025-12-09 15:47:55 +0000 UTC" firstStartedPulling="2025-12-09 15:47:58.953861352 +0000 UTC m=+973.273072021" lastFinishedPulling="2025-12-09 15:48:02.663221829 +0000 UTC m=+976.982432498" observedRunningTime="2025-12-09 15:48:04.283366763 +0000 UTC m=+978.602577442" watchObservedRunningTime="2025-12-09 15:48:04.291992808 +0000 UTC m=+978.611203477" Dec 09 15:48:04 crc kubenswrapper[4894]: I1209 15:48:04.306794 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=4.958527251 podStartE2EDuration="9.306775005s" podCreationTimestamp="2025-12-09 15:47:55 +0000 UTC" firstStartedPulling="2025-12-09 15:47:58.953877153 +0000 UTC m=+973.273087822" lastFinishedPulling="2025-12-09 15:48:03.302124897 +0000 UTC m=+977.621335576" observedRunningTime="2025-12-09 15:48:04.301747303 +0000 UTC m=+978.620957972" watchObservedRunningTime="2025-12-09 15:48:04.306775005 +0000 UTC m=+978.625985674" Dec 09 15:48:04 crc kubenswrapper[4894]: I1209 15:48:04.504743 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-11cf-account-create-update-ggv4k"] Dec 09 15:48:04 crc kubenswrapper[4894]: I1209 15:48:04.601110 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-hp2h5"] Dec 09 15:48:04 crc kubenswrapper[4894]: W1209 15:48:04.603091 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf6e6dd91_2ce9_4c86_8ea7_a06c42d84b0b.slice/crio-a354899b216e2e92fbf4a22f28a37624741c15faa2fd462462b45c9c42782c05 WatchSource:0}: Error finding container a354899b216e2e92fbf4a22f28a37624741c15faa2fd462462b45c9c42782c05: Status 404 returned error can't find the container with id a354899b216e2e92fbf4a22f28a37624741c15faa2fd462462b45c9c42782c05 Dec 09 15:48:05 crc kubenswrapper[4894]: I1209 15:48:05.239705 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Dec 09 15:48:05 crc kubenswrapper[4894]: I1209 15:48:05.270321 4894 generic.go:334] "Generic (PLEG): container finished" podID="f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b" containerID="763c224aa9948df6b7bf361a9d72c078324d66c43ccdf343d608de6643c86769" exitCode=0 Dec 09 15:48:05 crc kubenswrapper[4894]: I1209 15:48:05.270400 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-hp2h5" event={"ID":"f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b","Type":"ContainerDied","Data":"763c224aa9948df6b7bf361a9d72c078324d66c43ccdf343d608de6643c86769"} Dec 09 15:48:05 crc kubenswrapper[4894]: I1209 15:48:05.270428 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-hp2h5" event={"ID":"f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b","Type":"ContainerStarted","Data":"a354899b216e2e92fbf4a22f28a37624741c15faa2fd462462b45c9c42782c05"} Dec 09 15:48:05 crc kubenswrapper[4894]: I1209 15:48:05.272544 4894 generic.go:334] "Generic (PLEG): container finished" podID="9d962c43-b57b-4db6-bda2-c6b48bb7335a" containerID="9d4e6632f32fe729acf3df190912cbb2f04261a364e68d5f5cde2931b3df2593" exitCode=0 Dec 09 15:48:05 crc kubenswrapper[4894]: I1209 15:48:05.273966 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-11cf-account-create-update-ggv4k" event={"ID":"9d962c43-b57b-4db6-bda2-c6b48bb7335a","Type":"ContainerDied","Data":"9d4e6632f32fe729acf3df190912cbb2f04261a364e68d5f5cde2931b3df2593"} Dec 09 15:48:05 crc kubenswrapper[4894]: I1209 15:48:05.274027 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-11cf-account-create-update-ggv4k" event={"ID":"9d962c43-b57b-4db6-bda2-c6b48bb7335a","Type":"ContainerStarted","Data":"b88259faf20b64aa0f9e25f5d670c38596a6b97812c0b4d026434ae84096b581"} Dec 09 15:48:05 crc kubenswrapper[4894]: I1209 15:48:05.337720 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Dec 09 15:48:06 crc kubenswrapper[4894]: I1209 15:48:06.725837 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-11cf-account-create-update-ggv4k" Dec 09 15:48:06 crc kubenswrapper[4894]: I1209 15:48:06.732016 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hp2h5" Dec 09 15:48:06 crc kubenswrapper[4894]: I1209 15:48:06.844789 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d962c43-b57b-4db6-bda2-c6b48bb7335a-operator-scripts\") pod \"9d962c43-b57b-4db6-bda2-c6b48bb7335a\" (UID: \"9d962c43-b57b-4db6-bda2-c6b48bb7335a\") " Dec 09 15:48:06 crc kubenswrapper[4894]: I1209 15:48:06.845163 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b-operator-scripts\") pod \"f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b\" (UID: \"f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b\") " Dec 09 15:48:06 crc kubenswrapper[4894]: I1209 15:48:06.845201 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x8sfz\" (UniqueName: \"kubernetes.io/projected/f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b-kube-api-access-x8sfz\") pod \"f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b\" (UID: \"f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b\") " Dec 09 15:48:06 crc kubenswrapper[4894]: I1209 15:48:06.845285 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hx5kl\" (UniqueName: \"kubernetes.io/projected/9d962c43-b57b-4db6-bda2-c6b48bb7335a-kube-api-access-hx5kl\") pod \"9d962c43-b57b-4db6-bda2-c6b48bb7335a\" (UID: \"9d962c43-b57b-4db6-bda2-c6b48bb7335a\") " Dec 09 15:48:06 crc kubenswrapper[4894]: I1209 15:48:06.845625 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b" (UID: "f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:06 crc kubenswrapper[4894]: I1209 15:48:06.846278 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:06 crc kubenswrapper[4894]: I1209 15:48:06.846461 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d962c43-b57b-4db6-bda2-c6b48bb7335a-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9d962c43-b57b-4db6-bda2-c6b48bb7335a" (UID: "9d962c43-b57b-4db6-bda2-c6b48bb7335a"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:06 crc kubenswrapper[4894]: I1209 15:48:06.850958 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b-kube-api-access-x8sfz" (OuterVolumeSpecName: "kube-api-access-x8sfz") pod "f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b" (UID: "f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b"). InnerVolumeSpecName "kube-api-access-x8sfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:48:06 crc kubenswrapper[4894]: I1209 15:48:06.851013 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d962c43-b57b-4db6-bda2-c6b48bb7335a-kube-api-access-hx5kl" (OuterVolumeSpecName: "kube-api-access-hx5kl") pod "9d962c43-b57b-4db6-bda2-c6b48bb7335a" (UID: "9d962c43-b57b-4db6-bda2-c6b48bb7335a"). InnerVolumeSpecName "kube-api-access-hx5kl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:48:06 crc kubenswrapper[4894]: I1209 15:48:06.948800 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x8sfz\" (UniqueName: \"kubernetes.io/projected/f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b-kube-api-access-x8sfz\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:06 crc kubenswrapper[4894]: I1209 15:48:06.948829 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hx5kl\" (UniqueName: \"kubernetes.io/projected/9d962c43-b57b-4db6-bda2-c6b48bb7335a-kube-api-access-hx5kl\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:06 crc kubenswrapper[4894]: I1209 15:48:06.948839 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9d962c43-b57b-4db6-bda2-c6b48bb7335a-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:07 crc kubenswrapper[4894]: I1209 15:48:07.289185 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-hp2h5" event={"ID":"f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b","Type":"ContainerDied","Data":"a354899b216e2e92fbf4a22f28a37624741c15faa2fd462462b45c9c42782c05"} Dec 09 15:48:07 crc kubenswrapper[4894]: I1209 15:48:07.289225 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a354899b216e2e92fbf4a22f28a37624741c15faa2fd462462b45c9c42782c05" Dec 09 15:48:07 crc kubenswrapper[4894]: I1209 15:48:07.289278 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-hp2h5" Dec 09 15:48:07 crc kubenswrapper[4894]: I1209 15:48:07.292039 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-11cf-account-create-update-ggv4k" event={"ID":"9d962c43-b57b-4db6-bda2-c6b48bb7335a","Type":"ContainerDied","Data":"b88259faf20b64aa0f9e25f5d670c38596a6b97812c0b4d026434ae84096b581"} Dec 09 15:48:07 crc kubenswrapper[4894]: I1209 15:48:07.292093 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b88259faf20b64aa0f9e25f5d670c38596a6b97812c0b4d026434ae84096b581" Dec 09 15:48:07 crc kubenswrapper[4894]: I1209 15:48:07.292146 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-11cf-account-create-update-ggv4k" Dec 09 15:48:07 crc kubenswrapper[4894]: I1209 15:48:07.998479 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-9j5zq"] Dec 09 15:48:07 crc kubenswrapper[4894]: E1209 15:48:07.998942 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d962c43-b57b-4db6-bda2-c6b48bb7335a" containerName="mariadb-account-create-update" Dec 09 15:48:07 crc kubenswrapper[4894]: I1209 15:48:07.998965 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d962c43-b57b-4db6-bda2-c6b48bb7335a" containerName="mariadb-account-create-update" Dec 09 15:48:07 crc kubenswrapper[4894]: E1209 15:48:07.999019 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b" containerName="mariadb-database-create" Dec 09 15:48:07 crc kubenswrapper[4894]: I1209 15:48:07.999031 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b" containerName="mariadb-database-create" Dec 09 15:48:07 crc kubenswrapper[4894]: I1209 15:48:07.999311 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d962c43-b57b-4db6-bda2-c6b48bb7335a" containerName="mariadb-account-create-update" Dec 09 15:48:07 crc kubenswrapper[4894]: I1209 15:48:07.999347 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b" containerName="mariadb-database-create" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.000149 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9j5zq" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.009128 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-9j5zq"] Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.103342 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-8cc8-account-create-update-8f2ql"] Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.104680 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8cc8-account-create-update-8f2ql" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.107934 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.137456 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8cc8-account-create-update-8f2ql"] Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.167529 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05b5bde5-3e09-4764-b41f-0de53ad5207b-operator-scripts\") pod \"keystone-db-create-9j5zq\" (UID: \"05b5bde5-3e09-4764-b41f-0de53ad5207b\") " pod="openstack/keystone-db-create-9j5zq" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.167612 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jpd9\" (UniqueName: \"kubernetes.io/projected/05b5bde5-3e09-4764-b41f-0de53ad5207b-kube-api-access-2jpd9\") pod \"keystone-db-create-9j5zq\" (UID: \"05b5bde5-3e09-4764-b41f-0de53ad5207b\") " pod="openstack/keystone-db-create-9j5zq" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.268557 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b968921-706b-4a6d-a667-2cc3389710ef-operator-scripts\") pod \"keystone-8cc8-account-create-update-8f2ql\" (UID: \"9b968921-706b-4a6d-a667-2cc3389710ef\") " pod="openstack/keystone-8cc8-account-create-update-8f2ql" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.269109 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w4ff\" (UniqueName: \"kubernetes.io/projected/9b968921-706b-4a6d-a667-2cc3389710ef-kube-api-access-5w4ff\") pod \"keystone-8cc8-account-create-update-8f2ql\" (UID: \"9b968921-706b-4a6d-a667-2cc3389710ef\") " pod="openstack/keystone-8cc8-account-create-update-8f2ql" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.269232 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05b5bde5-3e09-4764-b41f-0de53ad5207b-operator-scripts\") pod \"keystone-db-create-9j5zq\" (UID: \"05b5bde5-3e09-4764-b41f-0de53ad5207b\") " pod="openstack/keystone-db-create-9j5zq" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.269430 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jpd9\" (UniqueName: \"kubernetes.io/projected/05b5bde5-3e09-4764-b41f-0de53ad5207b-kube-api-access-2jpd9\") pod \"keystone-db-create-9j5zq\" (UID: \"05b5bde5-3e09-4764-b41f-0de53ad5207b\") " pod="openstack/keystone-db-create-9j5zq" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.270033 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05b5bde5-3e09-4764-b41f-0de53ad5207b-operator-scripts\") pod \"keystone-db-create-9j5zq\" (UID: \"05b5bde5-3e09-4764-b41f-0de53ad5207b\") " pod="openstack/keystone-db-create-9j5zq" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.290257 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jpd9\" (UniqueName: \"kubernetes.io/projected/05b5bde5-3e09-4764-b41f-0de53ad5207b-kube-api-access-2jpd9\") pod \"keystone-db-create-9j5zq\" (UID: \"05b5bde5-3e09-4764-b41f-0de53ad5207b\") " pod="openstack/keystone-db-create-9j5zq" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.323567 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9j5zq" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.372336 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b968921-706b-4a6d-a667-2cc3389710ef-operator-scripts\") pod \"keystone-8cc8-account-create-update-8f2ql\" (UID: \"9b968921-706b-4a6d-a667-2cc3389710ef\") " pod="openstack/keystone-8cc8-account-create-update-8f2ql" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.372410 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w4ff\" (UniqueName: \"kubernetes.io/projected/9b968921-706b-4a6d-a667-2cc3389710ef-kube-api-access-5w4ff\") pod \"keystone-8cc8-account-create-update-8f2ql\" (UID: \"9b968921-706b-4a6d-a667-2cc3389710ef\") " pod="openstack/keystone-8cc8-account-create-update-8f2ql" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.373215 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b968921-706b-4a6d-a667-2cc3389710ef-operator-scripts\") pod \"keystone-8cc8-account-create-update-8f2ql\" (UID: \"9b968921-706b-4a6d-a667-2cc3389710ef\") " pod="openstack/keystone-8cc8-account-create-update-8f2ql" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.390581 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w4ff\" (UniqueName: \"kubernetes.io/projected/9b968921-706b-4a6d-a667-2cc3389710ef-kube-api-access-5w4ff\") pod \"keystone-8cc8-account-create-update-8f2ql\" (UID: \"9b968921-706b-4a6d-a667-2cc3389710ef\") " pod="openstack/keystone-8cc8-account-create-update-8f2ql" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.442960 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8cc8-account-create-update-8f2ql" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.493870 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-tqxhm"] Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.496186 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-tqxhm" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.506521 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-tqxhm"] Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.612649 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-1747-account-create-update-6fqrf"] Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.613829 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1747-account-create-update-6fqrf" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.616252 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.620401 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-1747-account-create-update-6fqrf"] Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.678127 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc5ff393-b30e-448b-a618-09a3b3162c97-operator-scripts\") pod \"placement-db-create-tqxhm\" (UID: \"dc5ff393-b30e-448b-a618-09a3b3162c97\") " pod="openstack/placement-db-create-tqxhm" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.678211 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfvsd\" (UniqueName: \"kubernetes.io/projected/dc5ff393-b30e-448b-a618-09a3b3162c97-kube-api-access-bfvsd\") pod \"placement-db-create-tqxhm\" (UID: \"dc5ff393-b30e-448b-a618-09a3b3162c97\") " pod="openstack/placement-db-create-tqxhm" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.779365 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-292gz\" (UniqueName: \"kubernetes.io/projected/372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1-kube-api-access-292gz\") pod \"placement-1747-account-create-update-6fqrf\" (UID: \"372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1\") " pod="openstack/placement-1747-account-create-update-6fqrf" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.779503 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc5ff393-b30e-448b-a618-09a3b3162c97-operator-scripts\") pod \"placement-db-create-tqxhm\" (UID: \"dc5ff393-b30e-448b-a618-09a3b3162c97\") " pod="openstack/placement-db-create-tqxhm" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.779568 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfvsd\" (UniqueName: \"kubernetes.io/projected/dc5ff393-b30e-448b-a618-09a3b3162c97-kube-api-access-bfvsd\") pod \"placement-db-create-tqxhm\" (UID: \"dc5ff393-b30e-448b-a618-09a3b3162c97\") " pod="openstack/placement-db-create-tqxhm" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.779610 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1-operator-scripts\") pod \"placement-1747-account-create-update-6fqrf\" (UID: \"372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1\") " pod="openstack/placement-1747-account-create-update-6fqrf" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.780315 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc5ff393-b30e-448b-a618-09a3b3162c97-operator-scripts\") pod \"placement-db-create-tqxhm\" (UID: \"dc5ff393-b30e-448b-a618-09a3b3162c97\") " pod="openstack/placement-db-create-tqxhm" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.801528 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfvsd\" (UniqueName: \"kubernetes.io/projected/dc5ff393-b30e-448b-a618-09a3b3162c97-kube-api-access-bfvsd\") pod \"placement-db-create-tqxhm\" (UID: \"dc5ff393-b30e-448b-a618-09a3b3162c97\") " pod="openstack/placement-db-create-tqxhm" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.804410 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-9j5zq"] Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.870571 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-tqxhm" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.880775 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1-operator-scripts\") pod \"placement-1747-account-create-update-6fqrf\" (UID: \"372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1\") " pod="openstack/placement-1747-account-create-update-6fqrf" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.880853 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-292gz\" (UniqueName: \"kubernetes.io/projected/372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1-kube-api-access-292gz\") pod \"placement-1747-account-create-update-6fqrf\" (UID: \"372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1\") " pod="openstack/placement-1747-account-create-update-6fqrf" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.881596 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1-operator-scripts\") pod \"placement-1747-account-create-update-6fqrf\" (UID: \"372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1\") " pod="openstack/placement-1747-account-create-update-6fqrf" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.908756 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-292gz\" (UniqueName: \"kubernetes.io/projected/372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1-kube-api-access-292gz\") pod \"placement-1747-account-create-update-6fqrf\" (UID: \"372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1\") " pod="openstack/placement-1747-account-create-update-6fqrf" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.918212 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-h74cz"] Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.919401 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-h74cz" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.927491 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-pj76v" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.931987 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.932485 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1747-account-create-update-6fqrf" Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.938860 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-h74cz"] Dec 09 15:48:08 crc kubenswrapper[4894]: W1209 15:48:08.972144 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9b968921_706b_4a6d_a667_2cc3389710ef.slice/crio-11beaab713522f18f2cf501e1233eb92e3c3d8b4515f559e13c03dae0c11e86b WatchSource:0}: Error finding container 11beaab713522f18f2cf501e1233eb92e3c3d8b4515f559e13c03dae0c11e86b: Status 404 returned error can't find the container with id 11beaab713522f18f2cf501e1233eb92e3c3d8b4515f559e13c03dae0c11e86b Dec 09 15:48:08 crc kubenswrapper[4894]: I1209 15:48:08.976875 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-8cc8-account-create-update-8f2ql"] Dec 09 15:48:09 crc kubenswrapper[4894]: I1209 15:48:09.085976 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26cc6eda-fee8-4183-b613-327cc2fad4b6-combined-ca-bundle\") pod \"glance-db-sync-h74cz\" (UID: \"26cc6eda-fee8-4183-b613-327cc2fad4b6\") " pod="openstack/glance-db-sync-h74cz" Dec 09 15:48:09 crc kubenswrapper[4894]: I1209 15:48:09.086578 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwcsb\" (UniqueName: \"kubernetes.io/projected/26cc6eda-fee8-4183-b613-327cc2fad4b6-kube-api-access-dwcsb\") pod \"glance-db-sync-h74cz\" (UID: \"26cc6eda-fee8-4183-b613-327cc2fad4b6\") " pod="openstack/glance-db-sync-h74cz" Dec 09 15:48:09 crc kubenswrapper[4894]: I1209 15:48:09.086611 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26cc6eda-fee8-4183-b613-327cc2fad4b6-config-data\") pod \"glance-db-sync-h74cz\" (UID: \"26cc6eda-fee8-4183-b613-327cc2fad4b6\") " pod="openstack/glance-db-sync-h74cz" Dec 09 15:48:09 crc kubenswrapper[4894]: I1209 15:48:09.086694 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/26cc6eda-fee8-4183-b613-327cc2fad4b6-db-sync-config-data\") pod \"glance-db-sync-h74cz\" (UID: \"26cc6eda-fee8-4183-b613-327cc2fad4b6\") " pod="openstack/glance-db-sync-h74cz" Dec 09 15:48:09 crc kubenswrapper[4894]: I1209 15:48:09.189490 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwcsb\" (UniqueName: \"kubernetes.io/projected/26cc6eda-fee8-4183-b613-327cc2fad4b6-kube-api-access-dwcsb\") pod \"glance-db-sync-h74cz\" (UID: \"26cc6eda-fee8-4183-b613-327cc2fad4b6\") " pod="openstack/glance-db-sync-h74cz" Dec 09 15:48:09 crc kubenswrapper[4894]: I1209 15:48:09.189553 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26cc6eda-fee8-4183-b613-327cc2fad4b6-config-data\") pod \"glance-db-sync-h74cz\" (UID: \"26cc6eda-fee8-4183-b613-327cc2fad4b6\") " pod="openstack/glance-db-sync-h74cz" Dec 09 15:48:09 crc kubenswrapper[4894]: I1209 15:48:09.189582 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/26cc6eda-fee8-4183-b613-327cc2fad4b6-db-sync-config-data\") pod \"glance-db-sync-h74cz\" (UID: \"26cc6eda-fee8-4183-b613-327cc2fad4b6\") " pod="openstack/glance-db-sync-h74cz" Dec 09 15:48:09 crc kubenswrapper[4894]: I1209 15:48:09.189749 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26cc6eda-fee8-4183-b613-327cc2fad4b6-combined-ca-bundle\") pod \"glance-db-sync-h74cz\" (UID: \"26cc6eda-fee8-4183-b613-327cc2fad4b6\") " pod="openstack/glance-db-sync-h74cz" Dec 09 15:48:09 crc kubenswrapper[4894]: I1209 15:48:09.198125 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/26cc6eda-fee8-4183-b613-327cc2fad4b6-db-sync-config-data\") pod \"glance-db-sync-h74cz\" (UID: \"26cc6eda-fee8-4183-b613-327cc2fad4b6\") " pod="openstack/glance-db-sync-h74cz" Dec 09 15:48:09 crc kubenswrapper[4894]: I1209 15:48:09.198333 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26cc6eda-fee8-4183-b613-327cc2fad4b6-combined-ca-bundle\") pod \"glance-db-sync-h74cz\" (UID: \"26cc6eda-fee8-4183-b613-327cc2fad4b6\") " pod="openstack/glance-db-sync-h74cz" Dec 09 15:48:09 crc kubenswrapper[4894]: I1209 15:48:09.200003 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26cc6eda-fee8-4183-b613-327cc2fad4b6-config-data\") pod \"glance-db-sync-h74cz\" (UID: \"26cc6eda-fee8-4183-b613-327cc2fad4b6\") " pod="openstack/glance-db-sync-h74cz" Dec 09 15:48:09 crc kubenswrapper[4894]: I1209 15:48:09.211633 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwcsb\" (UniqueName: \"kubernetes.io/projected/26cc6eda-fee8-4183-b613-327cc2fad4b6-kube-api-access-dwcsb\") pod \"glance-db-sync-h74cz\" (UID: \"26cc6eda-fee8-4183-b613-327cc2fad4b6\") " pod="openstack/glance-db-sync-h74cz" Dec 09 15:48:09 crc kubenswrapper[4894]: I1209 15:48:09.247861 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-h74cz" Dec 09 15:48:09 crc kubenswrapper[4894]: I1209 15:48:09.314285 4894 generic.go:334] "Generic (PLEG): container finished" podID="05b5bde5-3e09-4764-b41f-0de53ad5207b" containerID="fb9f558d8fc0848c266f1b70acc7ad89c12ec4368ceb418c4b5c2a60933c83ce" exitCode=0 Dec 09 15:48:09 crc kubenswrapper[4894]: I1209 15:48:09.314402 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9j5zq" event={"ID":"05b5bde5-3e09-4764-b41f-0de53ad5207b","Type":"ContainerDied","Data":"fb9f558d8fc0848c266f1b70acc7ad89c12ec4368ceb418c4b5c2a60933c83ce"} Dec 09 15:48:09 crc kubenswrapper[4894]: I1209 15:48:09.314435 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9j5zq" event={"ID":"05b5bde5-3e09-4764-b41f-0de53ad5207b","Type":"ContainerStarted","Data":"e2f5d30e1a3dbb2ae7abd9a42bc6c791ae2dc8047e01af0cd69fdc26dff8bb1b"} Dec 09 15:48:09 crc kubenswrapper[4894]: I1209 15:48:09.318770 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8cc8-account-create-update-8f2ql" event={"ID":"9b968921-706b-4a6d-a667-2cc3389710ef","Type":"ContainerStarted","Data":"da5a8a897a96863c3ab11ba7f22b4a58b681672b7e026cd0dc87fd9996e2d966"} Dec 09 15:48:09 crc kubenswrapper[4894]: I1209 15:48:09.318824 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8cc8-account-create-update-8f2ql" event={"ID":"9b968921-706b-4a6d-a667-2cc3389710ef","Type":"ContainerStarted","Data":"11beaab713522f18f2cf501e1233eb92e3c3d8b4515f559e13c03dae0c11e86b"} Dec 09 15:48:09 crc kubenswrapper[4894]: I1209 15:48:09.365001 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-8cc8-account-create-update-8f2ql" podStartSLOduration=1.364983885 podStartE2EDuration="1.364983885s" podCreationTimestamp="2025-12-09 15:48:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:48:09.3575086 +0000 UTC m=+983.676719269" watchObservedRunningTime="2025-12-09 15:48:09.364983885 +0000 UTC m=+983.684194554" Dec 09 15:48:09 crc kubenswrapper[4894]: I1209 15:48:09.392072 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-tqxhm"] Dec 09 15:48:09 crc kubenswrapper[4894]: W1209 15:48:09.426800 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddc5ff393_b30e_448b_a618_09a3b3162c97.slice/crio-b23fd545ae0e2b49860cc581835cbfc13ca358b020f6fd0d0fef0ba0c6b5744c WatchSource:0}: Error finding container b23fd545ae0e2b49860cc581835cbfc13ca358b020f6fd0d0fef0ba0c6b5744c: Status 404 returned error can't find the container with id b23fd545ae0e2b49860cc581835cbfc13ca358b020f6fd0d0fef0ba0c6b5744c Dec 09 15:48:09 crc kubenswrapper[4894]: I1209 15:48:09.473499 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-1747-account-create-update-6fqrf"] Dec 09 15:48:09 crc kubenswrapper[4894]: W1209 15:48:09.475621 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod372381bb_d5ff_47d7_a2bb_9ba8cfeacbe1.slice/crio-8b064052eefc651c04bddfc464ef21a66559cd47a31718d96f23932505b5f49a WatchSource:0}: Error finding container 8b064052eefc651c04bddfc464ef21a66559cd47a31718d96f23932505b5f49a: Status 404 returned error can't find the container with id 8b064052eefc651c04bddfc464ef21a66559cd47a31718d96f23932505b5f49a Dec 09 15:48:09 crc kubenswrapper[4894]: I1209 15:48:09.791374 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-h74cz"] Dec 09 15:48:10 crc kubenswrapper[4894]: I1209 15:48:10.331327 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-h74cz" event={"ID":"26cc6eda-fee8-4183-b613-327cc2fad4b6","Type":"ContainerStarted","Data":"cfc934393d82112faa6a48e6f594b5db47169f27af7006a38df31f1aabdbfa08"} Dec 09 15:48:10 crc kubenswrapper[4894]: I1209 15:48:10.333128 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1747-account-create-update-6fqrf" event={"ID":"372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1","Type":"ContainerStarted","Data":"e58e180c7e9f42cce467df5cf3d928383cd93d1b7626a554d1b4b50b83382cf7"} Dec 09 15:48:10 crc kubenswrapper[4894]: I1209 15:48:10.333161 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1747-account-create-update-6fqrf" event={"ID":"372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1","Type":"ContainerStarted","Data":"8b064052eefc651c04bddfc464ef21a66559cd47a31718d96f23932505b5f49a"} Dec 09 15:48:10 crc kubenswrapper[4894]: I1209 15:48:10.334625 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-tqxhm" event={"ID":"dc5ff393-b30e-448b-a618-09a3b3162c97","Type":"ContainerStarted","Data":"e8ff58696a142afbc97141c68c1505f7807ba4c434f4cc62d88c65a3afab76e5"} Dec 09 15:48:10 crc kubenswrapper[4894]: I1209 15:48:10.334667 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-tqxhm" event={"ID":"dc5ff393-b30e-448b-a618-09a3b3162c97","Type":"ContainerStarted","Data":"b23fd545ae0e2b49860cc581835cbfc13ca358b020f6fd0d0fef0ba0c6b5744c"} Dec 09 15:48:10 crc kubenswrapper[4894]: I1209 15:48:10.337390 4894 generic.go:334] "Generic (PLEG): container finished" podID="9b968921-706b-4a6d-a667-2cc3389710ef" containerID="da5a8a897a96863c3ab11ba7f22b4a58b681672b7e026cd0dc87fd9996e2d966" exitCode=0 Dec 09 15:48:10 crc kubenswrapper[4894]: I1209 15:48:10.337456 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8cc8-account-create-update-8f2ql" event={"ID":"9b968921-706b-4a6d-a667-2cc3389710ef","Type":"ContainerDied","Data":"da5a8a897a96863c3ab11ba7f22b4a58b681672b7e026cd0dc87fd9996e2d966"} Dec 09 15:48:10 crc kubenswrapper[4894]: I1209 15:48:10.355921 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-1747-account-create-update-6fqrf" podStartSLOduration=2.355896483 podStartE2EDuration="2.355896483s" podCreationTimestamp="2025-12-09 15:48:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:48:10.350749268 +0000 UTC m=+984.669959937" watchObservedRunningTime="2025-12-09 15:48:10.355896483 +0000 UTC m=+984.675107152" Dec 09 15:48:10 crc kubenswrapper[4894]: I1209 15:48:10.390677 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-create-tqxhm" podStartSLOduration=2.390659232 podStartE2EDuration="2.390659232s" podCreationTimestamp="2025-12-09 15:48:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:48:10.386834342 +0000 UTC m=+984.706045011" watchObservedRunningTime="2025-12-09 15:48:10.390659232 +0000 UTC m=+984.709869901" Dec 09 15:48:10 crc kubenswrapper[4894]: I1209 15:48:10.816748 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9j5zq" Dec 09 15:48:10 crc kubenswrapper[4894]: I1209 15:48:10.864185 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7878659675-bwbdn" Dec 09 15:48:10 crc kubenswrapper[4894]: I1209 15:48:10.921361 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05b5bde5-3e09-4764-b41f-0de53ad5207b-operator-scripts\") pod \"05b5bde5-3e09-4764-b41f-0de53ad5207b\" (UID: \"05b5bde5-3e09-4764-b41f-0de53ad5207b\") " Dec 09 15:48:10 crc kubenswrapper[4894]: I1209 15:48:10.921533 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2jpd9\" (UniqueName: \"kubernetes.io/projected/05b5bde5-3e09-4764-b41f-0de53ad5207b-kube-api-access-2jpd9\") pod \"05b5bde5-3e09-4764-b41f-0de53ad5207b\" (UID: \"05b5bde5-3e09-4764-b41f-0de53ad5207b\") " Dec 09 15:48:10 crc kubenswrapper[4894]: I1209 15:48:10.922140 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05b5bde5-3e09-4764-b41f-0de53ad5207b-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "05b5bde5-3e09-4764-b41f-0de53ad5207b" (UID: "05b5bde5-3e09-4764-b41f-0de53ad5207b"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:10 crc kubenswrapper[4894]: I1209 15:48:10.930494 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05b5bde5-3e09-4764-b41f-0de53ad5207b-kube-api-access-2jpd9" (OuterVolumeSpecName: "kube-api-access-2jpd9") pod "05b5bde5-3e09-4764-b41f-0de53ad5207b" (UID: "05b5bde5-3e09-4764-b41f-0de53ad5207b"). InnerVolumeSpecName "kube-api-access-2jpd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:48:11 crc kubenswrapper[4894]: I1209 15:48:11.024458 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/05b5bde5-3e09-4764-b41f-0de53ad5207b-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:11 crc kubenswrapper[4894]: I1209 15:48:11.024495 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2jpd9\" (UniqueName: \"kubernetes.io/projected/05b5bde5-3e09-4764-b41f-0de53ad5207b-kube-api-access-2jpd9\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:11 crc kubenswrapper[4894]: I1209 15:48:11.214921 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-586b989cdc-946tp" Dec 09 15:48:11 crc kubenswrapper[4894]: I1209 15:48:11.291977 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7878659675-bwbdn"] Dec 09 15:48:11 crc kubenswrapper[4894]: I1209 15:48:11.353584 4894 generic.go:334] "Generic (PLEG): container finished" podID="372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1" containerID="e58e180c7e9f42cce467df5cf3d928383cd93d1b7626a554d1b4b50b83382cf7" exitCode=0 Dec 09 15:48:11 crc kubenswrapper[4894]: I1209 15:48:11.353655 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1747-account-create-update-6fqrf" event={"ID":"372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1","Type":"ContainerDied","Data":"e58e180c7e9f42cce467df5cf3d928383cd93d1b7626a554d1b4b50b83382cf7"} Dec 09 15:48:11 crc kubenswrapper[4894]: I1209 15:48:11.357462 4894 generic.go:334] "Generic (PLEG): container finished" podID="dc5ff393-b30e-448b-a618-09a3b3162c97" containerID="e8ff58696a142afbc97141c68c1505f7807ba4c434f4cc62d88c65a3afab76e5" exitCode=0 Dec 09 15:48:11 crc kubenswrapper[4894]: I1209 15:48:11.357528 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-tqxhm" event={"ID":"dc5ff393-b30e-448b-a618-09a3b3162c97","Type":"ContainerDied","Data":"e8ff58696a142afbc97141c68c1505f7807ba4c434f4cc62d88c65a3afab76e5"} Dec 09 15:48:11 crc kubenswrapper[4894]: I1209 15:48:11.365085 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-9j5zq" Dec 09 15:48:11 crc kubenswrapper[4894]: I1209 15:48:11.365434 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-9j5zq" event={"ID":"05b5bde5-3e09-4764-b41f-0de53ad5207b","Type":"ContainerDied","Data":"e2f5d30e1a3dbb2ae7abd9a42bc6c791ae2dc8047e01af0cd69fdc26dff8bb1b"} Dec 09 15:48:11 crc kubenswrapper[4894]: I1209 15:48:11.365510 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2f5d30e1a3dbb2ae7abd9a42bc6c791ae2dc8047e01af0cd69fdc26dff8bb1b" Dec 09 15:48:11 crc kubenswrapper[4894]: I1209 15:48:11.366722 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7878659675-bwbdn" podUID="7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad" containerName="dnsmasq-dns" containerID="cri-o://aa3a6add204fac867d21e0c55ff60c6bf702cb003d34594c10dc1afc37493194" gracePeriod=10 Dec 09 15:48:11 crc kubenswrapper[4894]: I1209 15:48:11.731156 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8cc8-account-create-update-8f2ql" Dec 09 15:48:11 crc kubenswrapper[4894]: I1209 15:48:11.844788 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b968921-706b-4a6d-a667-2cc3389710ef-operator-scripts\") pod \"9b968921-706b-4a6d-a667-2cc3389710ef\" (UID: \"9b968921-706b-4a6d-a667-2cc3389710ef\") " Dec 09 15:48:11 crc kubenswrapper[4894]: I1209 15:48:11.844835 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5w4ff\" (UniqueName: \"kubernetes.io/projected/9b968921-706b-4a6d-a667-2cc3389710ef-kube-api-access-5w4ff\") pod \"9b968921-706b-4a6d-a667-2cc3389710ef\" (UID: \"9b968921-706b-4a6d-a667-2cc3389710ef\") " Dec 09 15:48:11 crc kubenswrapper[4894]: I1209 15:48:11.845305 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9b968921-706b-4a6d-a667-2cc3389710ef-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "9b968921-706b-4a6d-a667-2cc3389710ef" (UID: "9b968921-706b-4a6d-a667-2cc3389710ef"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:11 crc kubenswrapper[4894]: I1209 15:48:11.849390 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9b968921-706b-4a6d-a667-2cc3389710ef-kube-api-access-5w4ff" (OuterVolumeSpecName: "kube-api-access-5w4ff") pod "9b968921-706b-4a6d-a667-2cc3389710ef" (UID: "9b968921-706b-4a6d-a667-2cc3389710ef"). InnerVolumeSpecName "kube-api-access-5w4ff". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:48:11 crc kubenswrapper[4894]: I1209 15:48:11.918247 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7878659675-bwbdn" Dec 09 15:48:11 crc kubenswrapper[4894]: I1209 15:48:11.947861 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/9b968921-706b-4a6d-a667-2cc3389710ef-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:11 crc kubenswrapper[4894]: I1209 15:48:11.948006 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5w4ff\" (UniqueName: \"kubernetes.io/projected/9b968921-706b-4a6d-a667-2cc3389710ef-kube-api-access-5w4ff\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.049211 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-dns-svc\") pod \"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad\" (UID: \"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad\") " Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.050072 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9zs7j\" (UniqueName: \"kubernetes.io/projected/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-kube-api-access-9zs7j\") pod \"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad\" (UID: \"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad\") " Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.050131 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-ovsdbserver-nb\") pod \"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad\" (UID: \"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad\") " Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.050178 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-config\") pod \"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad\" (UID: \"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad\") " Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.054788 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-kube-api-access-9zs7j" (OuterVolumeSpecName: "kube-api-access-9zs7j") pod "7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad" (UID: "7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad"). InnerVolumeSpecName "kube-api-access-9zs7j". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.092379 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad" (UID: "7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.096257 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-config" (OuterVolumeSpecName: "config") pod "7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad" (UID: "7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.097901 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad" (UID: "7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.154612 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9zs7j\" (UniqueName: \"kubernetes.io/projected/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-kube-api-access-9zs7j\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.154673 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.154686 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.154694 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.378853 4894 generic.go:334] "Generic (PLEG): container finished" podID="7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad" containerID="aa3a6add204fac867d21e0c55ff60c6bf702cb003d34594c10dc1afc37493194" exitCode=0 Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.378948 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7878659675-bwbdn" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.378961 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7878659675-bwbdn" event={"ID":"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad","Type":"ContainerDied","Data":"aa3a6add204fac867d21e0c55ff60c6bf702cb003d34594c10dc1afc37493194"} Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.379010 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7878659675-bwbdn" event={"ID":"7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad","Type":"ContainerDied","Data":"a0fdadc383b2f2b90bcded1445ddccf8bfd18a72e2ebe60207251ee012fc1455"} Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.379038 4894 scope.go:117] "RemoveContainer" containerID="aa3a6add204fac867d21e0c55ff60c6bf702cb003d34594c10dc1afc37493194" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.383011 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-8cc8-account-create-update-8f2ql" event={"ID":"9b968921-706b-4a6d-a667-2cc3389710ef","Type":"ContainerDied","Data":"11beaab713522f18f2cf501e1233eb92e3c3d8b4515f559e13c03dae0c11e86b"} Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.383059 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="11beaab713522f18f2cf501e1233eb92e3c3d8b4515f559e13c03dae0c11e86b" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.383060 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-8cc8-account-create-update-8f2ql" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.419410 4894 scope.go:117] "RemoveContainer" containerID="bc6f9ffb28b4541dc62860082365a7be6cc5d642dc81599c8989a128bc35624c" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.429905 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7878659675-bwbdn"] Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.443229 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7878659675-bwbdn"] Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.461257 4894 scope.go:117] "RemoveContainer" containerID="aa3a6add204fac867d21e0c55ff60c6bf702cb003d34594c10dc1afc37493194" Dec 09 15:48:12 crc kubenswrapper[4894]: E1209 15:48:12.461735 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa3a6add204fac867d21e0c55ff60c6bf702cb003d34594c10dc1afc37493194\": container with ID starting with aa3a6add204fac867d21e0c55ff60c6bf702cb003d34594c10dc1afc37493194 not found: ID does not exist" containerID="aa3a6add204fac867d21e0c55ff60c6bf702cb003d34594c10dc1afc37493194" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.461877 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa3a6add204fac867d21e0c55ff60c6bf702cb003d34594c10dc1afc37493194"} err="failed to get container status \"aa3a6add204fac867d21e0c55ff60c6bf702cb003d34594c10dc1afc37493194\": rpc error: code = NotFound desc = could not find container \"aa3a6add204fac867d21e0c55ff60c6bf702cb003d34594c10dc1afc37493194\": container with ID starting with aa3a6add204fac867d21e0c55ff60c6bf702cb003d34594c10dc1afc37493194 not found: ID does not exist" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.462019 4894 scope.go:117] "RemoveContainer" containerID="bc6f9ffb28b4541dc62860082365a7be6cc5d642dc81599c8989a128bc35624c" Dec 09 15:48:12 crc kubenswrapper[4894]: E1209 15:48:12.463795 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bc6f9ffb28b4541dc62860082365a7be6cc5d642dc81599c8989a128bc35624c\": container with ID starting with bc6f9ffb28b4541dc62860082365a7be6cc5d642dc81599c8989a128bc35624c not found: ID does not exist" containerID="bc6f9ffb28b4541dc62860082365a7be6cc5d642dc81599c8989a128bc35624c" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.463842 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bc6f9ffb28b4541dc62860082365a7be6cc5d642dc81599c8989a128bc35624c"} err="failed to get container status \"bc6f9ffb28b4541dc62860082365a7be6cc5d642dc81599c8989a128bc35624c\": rpc error: code = NotFound desc = could not find container \"bc6f9ffb28b4541dc62860082365a7be6cc5d642dc81599c8989a128bc35624c\": container with ID starting with bc6f9ffb28b4541dc62860082365a7be6cc5d642dc81599c8989a128bc35624c not found: ID does not exist" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.841546 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-tqxhm" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.848736 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1747-account-create-update-6fqrf" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.987072 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-292gz\" (UniqueName: \"kubernetes.io/projected/372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1-kube-api-access-292gz\") pod \"372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1\" (UID: \"372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1\") " Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.987119 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc5ff393-b30e-448b-a618-09a3b3162c97-operator-scripts\") pod \"dc5ff393-b30e-448b-a618-09a3b3162c97\" (UID: \"dc5ff393-b30e-448b-a618-09a3b3162c97\") " Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.987144 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bfvsd\" (UniqueName: \"kubernetes.io/projected/dc5ff393-b30e-448b-a618-09a3b3162c97-kube-api-access-bfvsd\") pod \"dc5ff393-b30e-448b-a618-09a3b3162c97\" (UID: \"dc5ff393-b30e-448b-a618-09a3b3162c97\") " Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.987259 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1-operator-scripts\") pod \"372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1\" (UID: \"372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1\") " Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.988174 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1" (UID: "372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.988174 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dc5ff393-b30e-448b-a618-09a3b3162c97-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "dc5ff393-b30e-448b-a618-09a3b3162c97" (UID: "dc5ff393-b30e-448b-a618-09a3b3162c97"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.993264 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dc5ff393-b30e-448b-a618-09a3b3162c97-kube-api-access-bfvsd" (OuterVolumeSpecName: "kube-api-access-bfvsd") pod "dc5ff393-b30e-448b-a618-09a3b3162c97" (UID: "dc5ff393-b30e-448b-a618-09a3b3162c97"). InnerVolumeSpecName "kube-api-access-bfvsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:48:12 crc kubenswrapper[4894]: I1209 15:48:12.993446 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1-kube-api-access-292gz" (OuterVolumeSpecName: "kube-api-access-292gz") pod "372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1" (UID: "372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1"). InnerVolumeSpecName "kube-api-access-292gz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:48:13 crc kubenswrapper[4894]: I1209 15:48:13.089548 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:13 crc kubenswrapper[4894]: I1209 15:48:13.089603 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-292gz\" (UniqueName: \"kubernetes.io/projected/372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1-kube-api-access-292gz\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:13 crc kubenswrapper[4894]: I1209 15:48:13.089619 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/dc5ff393-b30e-448b-a618-09a3b3162c97-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:13 crc kubenswrapper[4894]: I1209 15:48:13.089652 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bfvsd\" (UniqueName: \"kubernetes.io/projected/dc5ff393-b30e-448b-a618-09a3b3162c97-kube-api-access-bfvsd\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:13 crc kubenswrapper[4894]: I1209 15:48:13.393659 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-1747-account-create-update-6fqrf" event={"ID":"372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1","Type":"ContainerDied","Data":"8b064052eefc651c04bddfc464ef21a66559cd47a31718d96f23932505b5f49a"} Dec 09 15:48:13 crc kubenswrapper[4894]: I1209 15:48:13.393746 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b064052eefc651c04bddfc464ef21a66559cd47a31718d96f23932505b5f49a" Dec 09 15:48:13 crc kubenswrapper[4894]: I1209 15:48:13.393826 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-1747-account-create-update-6fqrf" Dec 09 15:48:13 crc kubenswrapper[4894]: I1209 15:48:13.396605 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-tqxhm" event={"ID":"dc5ff393-b30e-448b-a618-09a3b3162c97","Type":"ContainerDied","Data":"b23fd545ae0e2b49860cc581835cbfc13ca358b020f6fd0d0fef0ba0c6b5744c"} Dec 09 15:48:13 crc kubenswrapper[4894]: I1209 15:48:13.396656 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b23fd545ae0e2b49860cc581835cbfc13ca358b020f6fd0d0fef0ba0c6b5744c" Dec 09 15:48:13 crc kubenswrapper[4894]: I1209 15:48:13.396723 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-tqxhm" Dec 09 15:48:14 crc kubenswrapper[4894]: I1209 15:48:14.119999 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad" path="/var/lib/kubelet/pods/7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad/volumes" Dec 09 15:48:16 crc kubenswrapper[4894]: I1209 15:48:16.061560 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Dec 09 15:48:18 crc kubenswrapper[4894]: I1209 15:48:18.438736 4894 generic.go:334] "Generic (PLEG): container finished" podID="493425e2-b4c7-4a1b-af05-44b4f44bd2b3" containerID="e0cc90efd20a8df80e1313ad14be34aed239c406e39225fc092a39ec293bc945" exitCode=0 Dec 09 15:48:18 crc kubenswrapper[4894]: I1209 15:48:18.438824 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"493425e2-b4c7-4a1b-af05-44b4f44bd2b3","Type":"ContainerDied","Data":"e0cc90efd20a8df80e1313ad14be34aed239c406e39225fc092a39ec293bc945"} Dec 09 15:48:18 crc kubenswrapper[4894]: I1209 15:48:18.443651 4894 generic.go:334] "Generic (PLEG): container finished" podID="44fdaebd-a96c-47f9-a421-a0984f78972e" containerID="89642012cdb0862001c73aa935177e0ea6e93168b497fc358d09b0a6cbda4b1a" exitCode=0 Dec 09 15:48:18 crc kubenswrapper[4894]: I1209 15:48:18.443689 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"44fdaebd-a96c-47f9-a421-a0984f78972e","Type":"ContainerDied","Data":"89642012cdb0862001c73aa935177e0ea6e93168b497fc358d09b0a6cbda4b1a"} Dec 09 15:48:19 crc kubenswrapper[4894]: I1209 15:48:19.640581 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-vk5h6" podUID="cf53d7df-bcc4-46b3-b874-dac48eca5bc4" containerName="ovn-controller" probeResult="failure" output=< Dec 09 15:48:19 crc kubenswrapper[4894]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 09 15:48:19 crc kubenswrapper[4894]: > Dec 09 15:48:22 crc kubenswrapper[4894]: I1209 15:48:22.482261 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-h74cz" event={"ID":"26cc6eda-fee8-4183-b613-327cc2fad4b6","Type":"ContainerStarted","Data":"a441a508256e662aad253db351dca851a1e46457e2b20ce5554b322ae16ffadc"} Dec 09 15:48:22 crc kubenswrapper[4894]: I1209 15:48:22.486575 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"493425e2-b4c7-4a1b-af05-44b4f44bd2b3","Type":"ContainerStarted","Data":"521f1e5dfce4a639e4244abf43400cbc86ec270828894f24b2104650d695e1f3"} Dec 09 15:48:22 crc kubenswrapper[4894]: I1209 15:48:22.487260 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 09 15:48:22 crc kubenswrapper[4894]: I1209 15:48:22.489669 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"44fdaebd-a96c-47f9-a421-a0984f78972e","Type":"ContainerStarted","Data":"0ebd4954c8e209f4c21212c42fd48068abec230259a29738503745230bd00e15"} Dec 09 15:48:22 crc kubenswrapper[4894]: I1209 15:48:22.489873 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:48:22 crc kubenswrapper[4894]: I1209 15:48:22.508027 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-h74cz" podStartSLOduration=3.045709961 podStartE2EDuration="14.50800452s" podCreationTimestamp="2025-12-09 15:48:08 +0000 UTC" firstStartedPulling="2025-12-09 15:48:09.808794844 +0000 UTC m=+984.128005513" lastFinishedPulling="2025-12-09 15:48:21.271089343 +0000 UTC m=+995.590300072" observedRunningTime="2025-12-09 15:48:22.503961684 +0000 UTC m=+996.823172353" watchObservedRunningTime="2025-12-09 15:48:22.50800452 +0000 UTC m=+996.827215189" Dec 09 15:48:22 crc kubenswrapper[4894]: I1209 15:48:22.537931 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=40.404938512 podStartE2EDuration="1m8.537909342s" podCreationTimestamp="2025-12-09 15:47:14 +0000 UTC" firstStartedPulling="2025-12-09 15:47:16.121998801 +0000 UTC m=+930.441209470" lastFinishedPulling="2025-12-09 15:47:44.254969631 +0000 UTC m=+958.574180300" observedRunningTime="2025-12-09 15:48:22.534506183 +0000 UTC m=+996.853716852" watchObservedRunningTime="2025-12-09 15:48:22.537909342 +0000 UTC m=+996.857120031" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.650286 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-vk5h6" podUID="cf53d7df-bcc4-46b3-b874-dac48eca5bc4" containerName="ovn-controller" probeResult="failure" output=< Dec 09 15:48:24 crc kubenswrapper[4894]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Dec 09 15:48:24 crc kubenswrapper[4894]: > Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.661522 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.665195 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-z7vcg" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.684216 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=43.366745511 podStartE2EDuration="1m11.684198516s" podCreationTimestamp="2025-12-09 15:47:13 +0000 UTC" firstStartedPulling="2025-12-09 15:47:15.936944451 +0000 UTC m=+930.256155120" lastFinishedPulling="2025-12-09 15:47:44.254397456 +0000 UTC m=+958.573608125" observedRunningTime="2025-12-09 15:48:22.576681735 +0000 UTC m=+996.895892424" watchObservedRunningTime="2025-12-09 15:48:24.684198516 +0000 UTC m=+999.003409185" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.880772 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-vk5h6-config-nr7kn"] Dec 09 15:48:24 crc kubenswrapper[4894]: E1209 15:48:24.881154 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad" containerName="dnsmasq-dns" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.881176 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad" containerName="dnsmasq-dns" Dec 09 15:48:24 crc kubenswrapper[4894]: E1209 15:48:24.881193 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dc5ff393-b30e-448b-a618-09a3b3162c97" containerName="mariadb-database-create" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.881202 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="dc5ff393-b30e-448b-a618-09a3b3162c97" containerName="mariadb-database-create" Dec 09 15:48:24 crc kubenswrapper[4894]: E1209 15:48:24.881228 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad" containerName="init" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.881240 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad" containerName="init" Dec 09 15:48:24 crc kubenswrapper[4894]: E1209 15:48:24.881256 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9b968921-706b-4a6d-a667-2cc3389710ef" containerName="mariadb-account-create-update" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.881264 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9b968921-706b-4a6d-a667-2cc3389710ef" containerName="mariadb-account-create-update" Dec 09 15:48:24 crc kubenswrapper[4894]: E1209 15:48:24.881282 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1" containerName="mariadb-account-create-update" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.881290 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1" containerName="mariadb-account-create-update" Dec 09 15:48:24 crc kubenswrapper[4894]: E1209 15:48:24.881303 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05b5bde5-3e09-4764-b41f-0de53ad5207b" containerName="mariadb-database-create" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.881332 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="05b5bde5-3e09-4764-b41f-0de53ad5207b" containerName="mariadb-database-create" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.881508 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="dc5ff393-b30e-448b-a618-09a3b3162c97" containerName="mariadb-database-create" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.881526 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9b968921-706b-4a6d-a667-2cc3389710ef" containerName="mariadb-account-create-update" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.881539 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1" containerName="mariadb-account-create-update" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.881560 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="05b5bde5-3e09-4764-b41f-0de53ad5207b" containerName="mariadb-database-create" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.881568 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dfad8b3-cbaa-4a80-ae85-799b1fa7f7ad" containerName="dnsmasq-dns" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.882180 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vk5h6-config-nr7kn" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.884289 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.909425 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vk5h6-config-nr7kn"] Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.987569 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8wdf\" (UniqueName: \"kubernetes.io/projected/1e8a9251-368a-47a4-872d-da9c17286dd4-kube-api-access-s8wdf\") pod \"ovn-controller-vk5h6-config-nr7kn\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " pod="openstack/ovn-controller-vk5h6-config-nr7kn" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.987626 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1e8a9251-368a-47a4-872d-da9c17286dd4-scripts\") pod \"ovn-controller-vk5h6-config-nr7kn\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " pod="openstack/ovn-controller-vk5h6-config-nr7kn" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.987672 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1e8a9251-368a-47a4-872d-da9c17286dd4-var-log-ovn\") pod \"ovn-controller-vk5h6-config-nr7kn\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " pod="openstack/ovn-controller-vk5h6-config-nr7kn" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.987741 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1e8a9251-368a-47a4-872d-da9c17286dd4-additional-scripts\") pod \"ovn-controller-vk5h6-config-nr7kn\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " pod="openstack/ovn-controller-vk5h6-config-nr7kn" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.987877 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1e8a9251-368a-47a4-872d-da9c17286dd4-var-run\") pod \"ovn-controller-vk5h6-config-nr7kn\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " pod="openstack/ovn-controller-vk5h6-config-nr7kn" Dec 09 15:48:24 crc kubenswrapper[4894]: I1209 15:48:24.987944 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1e8a9251-368a-47a4-872d-da9c17286dd4-var-run-ovn\") pod \"ovn-controller-vk5h6-config-nr7kn\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " pod="openstack/ovn-controller-vk5h6-config-nr7kn" Dec 09 15:48:25 crc kubenswrapper[4894]: I1209 15:48:25.088960 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1e8a9251-368a-47a4-872d-da9c17286dd4-scripts\") pod \"ovn-controller-vk5h6-config-nr7kn\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " pod="openstack/ovn-controller-vk5h6-config-nr7kn" Dec 09 15:48:25 crc kubenswrapper[4894]: I1209 15:48:25.089007 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1e8a9251-368a-47a4-872d-da9c17286dd4-var-log-ovn\") pod \"ovn-controller-vk5h6-config-nr7kn\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " pod="openstack/ovn-controller-vk5h6-config-nr7kn" Dec 09 15:48:25 crc kubenswrapper[4894]: I1209 15:48:25.089065 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1e8a9251-368a-47a4-872d-da9c17286dd4-additional-scripts\") pod \"ovn-controller-vk5h6-config-nr7kn\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " pod="openstack/ovn-controller-vk5h6-config-nr7kn" Dec 09 15:48:25 crc kubenswrapper[4894]: I1209 15:48:25.089105 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1e8a9251-368a-47a4-872d-da9c17286dd4-var-run\") pod \"ovn-controller-vk5h6-config-nr7kn\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " pod="openstack/ovn-controller-vk5h6-config-nr7kn" Dec 09 15:48:25 crc kubenswrapper[4894]: I1209 15:48:25.089122 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1e8a9251-368a-47a4-872d-da9c17286dd4-var-run-ovn\") pod \"ovn-controller-vk5h6-config-nr7kn\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " pod="openstack/ovn-controller-vk5h6-config-nr7kn" Dec 09 15:48:25 crc kubenswrapper[4894]: I1209 15:48:25.089182 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8wdf\" (UniqueName: \"kubernetes.io/projected/1e8a9251-368a-47a4-872d-da9c17286dd4-kube-api-access-s8wdf\") pod \"ovn-controller-vk5h6-config-nr7kn\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " pod="openstack/ovn-controller-vk5h6-config-nr7kn" Dec 09 15:48:25 crc kubenswrapper[4894]: I1209 15:48:25.091728 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1e8a9251-368a-47a4-872d-da9c17286dd4-scripts\") pod \"ovn-controller-vk5h6-config-nr7kn\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " pod="openstack/ovn-controller-vk5h6-config-nr7kn" Dec 09 15:48:25 crc kubenswrapper[4894]: I1209 15:48:25.091955 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1e8a9251-368a-47a4-872d-da9c17286dd4-var-log-ovn\") pod \"ovn-controller-vk5h6-config-nr7kn\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " pod="openstack/ovn-controller-vk5h6-config-nr7kn" Dec 09 15:48:25 crc kubenswrapper[4894]: I1209 15:48:25.091990 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1e8a9251-368a-47a4-872d-da9c17286dd4-var-run-ovn\") pod \"ovn-controller-vk5h6-config-nr7kn\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " pod="openstack/ovn-controller-vk5h6-config-nr7kn" Dec 09 15:48:25 crc kubenswrapper[4894]: I1209 15:48:25.092070 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1e8a9251-368a-47a4-872d-da9c17286dd4-var-run\") pod \"ovn-controller-vk5h6-config-nr7kn\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " pod="openstack/ovn-controller-vk5h6-config-nr7kn" Dec 09 15:48:25 crc kubenswrapper[4894]: I1209 15:48:25.092386 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1e8a9251-368a-47a4-872d-da9c17286dd4-additional-scripts\") pod \"ovn-controller-vk5h6-config-nr7kn\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " pod="openstack/ovn-controller-vk5h6-config-nr7kn" Dec 09 15:48:25 crc kubenswrapper[4894]: I1209 15:48:25.110834 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8wdf\" (UniqueName: \"kubernetes.io/projected/1e8a9251-368a-47a4-872d-da9c17286dd4-kube-api-access-s8wdf\") pod \"ovn-controller-vk5h6-config-nr7kn\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " pod="openstack/ovn-controller-vk5h6-config-nr7kn" Dec 09 15:48:25 crc kubenswrapper[4894]: I1209 15:48:25.205110 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vk5h6-config-nr7kn" Dec 09 15:48:25 crc kubenswrapper[4894]: I1209 15:48:25.661945 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vk5h6-config-nr7kn"] Dec 09 15:48:25 crc kubenswrapper[4894]: W1209 15:48:25.664785 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1e8a9251_368a_47a4_872d_da9c17286dd4.slice/crio-1af31c4bdbd73a213f5af95aeac4ff1b1a8599265016891711add5723ff77cc2 WatchSource:0}: Error finding container 1af31c4bdbd73a213f5af95aeac4ff1b1a8599265016891711add5723ff77cc2: Status 404 returned error can't find the container with id 1af31c4bdbd73a213f5af95aeac4ff1b1a8599265016891711add5723ff77cc2 Dec 09 15:48:26 crc kubenswrapper[4894]: I1209 15:48:26.522890 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vk5h6-config-nr7kn" event={"ID":"1e8a9251-368a-47a4-872d-da9c17286dd4","Type":"ContainerStarted","Data":"8127de34e5e54c1b1c5977a3a89313547d28de1afb989a55309d38e2e6488467"} Dec 09 15:48:26 crc kubenswrapper[4894]: I1209 15:48:26.523158 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vk5h6-config-nr7kn" event={"ID":"1e8a9251-368a-47a4-872d-da9c17286dd4","Type":"ContainerStarted","Data":"1af31c4bdbd73a213f5af95aeac4ff1b1a8599265016891711add5723ff77cc2"} Dec 09 15:48:27 crc kubenswrapper[4894]: I1209 15:48:27.537116 4894 generic.go:334] "Generic (PLEG): container finished" podID="1e8a9251-368a-47a4-872d-da9c17286dd4" containerID="8127de34e5e54c1b1c5977a3a89313547d28de1afb989a55309d38e2e6488467" exitCode=0 Dec 09 15:48:27 crc kubenswrapper[4894]: I1209 15:48:27.537234 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vk5h6-config-nr7kn" event={"ID":"1e8a9251-368a-47a4-872d-da9c17286dd4","Type":"ContainerDied","Data":"8127de34e5e54c1b1c5977a3a89313547d28de1afb989a55309d38e2e6488467"} Dec 09 15:48:28 crc kubenswrapper[4894]: I1209 15:48:28.830375 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vk5h6-config-nr7kn" Dec 09 15:48:28 crc kubenswrapper[4894]: I1209 15:48:28.959909 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1e8a9251-368a-47a4-872d-da9c17286dd4-additional-scripts\") pod \"1e8a9251-368a-47a4-872d-da9c17286dd4\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " Dec 09 15:48:28 crc kubenswrapper[4894]: I1209 15:48:28.960850 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e8a9251-368a-47a4-872d-da9c17286dd4-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "1e8a9251-368a-47a4-872d-da9c17286dd4" (UID: "1e8a9251-368a-47a4-872d-da9c17286dd4"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:28 crc kubenswrapper[4894]: I1209 15:48:28.960958 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1e8a9251-368a-47a4-872d-da9c17286dd4-scripts\") pod \"1e8a9251-368a-47a4-872d-da9c17286dd4\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " Dec 09 15:48:28 crc kubenswrapper[4894]: I1209 15:48:28.961851 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1e8a9251-368a-47a4-872d-da9c17286dd4-scripts" (OuterVolumeSpecName: "scripts") pod "1e8a9251-368a-47a4-872d-da9c17286dd4" (UID: "1e8a9251-368a-47a4-872d-da9c17286dd4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:28 crc kubenswrapper[4894]: I1209 15:48:28.961944 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1e8a9251-368a-47a4-872d-da9c17286dd4-var-log-ovn\") pod \"1e8a9251-368a-47a4-872d-da9c17286dd4\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " Dec 09 15:48:28 crc kubenswrapper[4894]: I1209 15:48:28.962018 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e8a9251-368a-47a4-872d-da9c17286dd4-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "1e8a9251-368a-47a4-872d-da9c17286dd4" (UID: "1e8a9251-368a-47a4-872d-da9c17286dd4"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:48:28 crc kubenswrapper[4894]: I1209 15:48:28.962086 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s8wdf\" (UniqueName: \"kubernetes.io/projected/1e8a9251-368a-47a4-872d-da9c17286dd4-kube-api-access-s8wdf\") pod \"1e8a9251-368a-47a4-872d-da9c17286dd4\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " Dec 09 15:48:28 crc kubenswrapper[4894]: I1209 15:48:28.962284 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e8a9251-368a-47a4-872d-da9c17286dd4-var-run" (OuterVolumeSpecName: "var-run") pod "1e8a9251-368a-47a4-872d-da9c17286dd4" (UID: "1e8a9251-368a-47a4-872d-da9c17286dd4"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:48:28 crc kubenswrapper[4894]: I1209 15:48:28.962646 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1e8a9251-368a-47a4-872d-da9c17286dd4-var-run\") pod \"1e8a9251-368a-47a4-872d-da9c17286dd4\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " Dec 09 15:48:28 crc kubenswrapper[4894]: I1209 15:48:28.962709 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1e8a9251-368a-47a4-872d-da9c17286dd4-var-run-ovn\") pod \"1e8a9251-368a-47a4-872d-da9c17286dd4\" (UID: \"1e8a9251-368a-47a4-872d-da9c17286dd4\") " Dec 09 15:48:28 crc kubenswrapper[4894]: I1209 15:48:28.962807 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1e8a9251-368a-47a4-872d-da9c17286dd4-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "1e8a9251-368a-47a4-872d-da9c17286dd4" (UID: "1e8a9251-368a-47a4-872d-da9c17286dd4"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:48:28 crc kubenswrapper[4894]: I1209 15:48:28.963171 4894 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/1e8a9251-368a-47a4-872d-da9c17286dd4-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:28 crc kubenswrapper[4894]: I1209 15:48:28.963189 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1e8a9251-368a-47a4-872d-da9c17286dd4-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:28 crc kubenswrapper[4894]: I1209 15:48:28.963197 4894 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/1e8a9251-368a-47a4-872d-da9c17286dd4-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:28 crc kubenswrapper[4894]: I1209 15:48:28.963207 4894 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/1e8a9251-368a-47a4-872d-da9c17286dd4-var-run\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:28 crc kubenswrapper[4894]: I1209 15:48:28.963216 4894 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/1e8a9251-368a-47a4-872d-da9c17286dd4-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:28 crc kubenswrapper[4894]: I1209 15:48:28.967332 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e8a9251-368a-47a4-872d-da9c17286dd4-kube-api-access-s8wdf" (OuterVolumeSpecName: "kube-api-access-s8wdf") pod "1e8a9251-368a-47a4-872d-da9c17286dd4" (UID: "1e8a9251-368a-47a4-872d-da9c17286dd4"). InnerVolumeSpecName "kube-api-access-s8wdf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:48:29 crc kubenswrapper[4894]: I1209 15:48:29.064944 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s8wdf\" (UniqueName: \"kubernetes.io/projected/1e8a9251-368a-47a4-872d-da9c17286dd4-kube-api-access-s8wdf\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:29 crc kubenswrapper[4894]: I1209 15:48:29.552821 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vk5h6-config-nr7kn" event={"ID":"1e8a9251-368a-47a4-872d-da9c17286dd4","Type":"ContainerDied","Data":"1af31c4bdbd73a213f5af95aeac4ff1b1a8599265016891711add5723ff77cc2"} Dec 09 15:48:29 crc kubenswrapper[4894]: I1209 15:48:29.552862 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1af31c4bdbd73a213f5af95aeac4ff1b1a8599265016891711add5723ff77cc2" Dec 09 15:48:29 crc kubenswrapper[4894]: I1209 15:48:29.552891 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vk5h6-config-nr7kn" Dec 09 15:48:29 crc kubenswrapper[4894]: I1209 15:48:29.554184 4894 generic.go:334] "Generic (PLEG): container finished" podID="26cc6eda-fee8-4183-b613-327cc2fad4b6" containerID="a441a508256e662aad253db351dca851a1e46457e2b20ce5554b322ae16ffadc" exitCode=0 Dec 09 15:48:29 crc kubenswrapper[4894]: I1209 15:48:29.554213 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-h74cz" event={"ID":"26cc6eda-fee8-4183-b613-327cc2fad4b6","Type":"ContainerDied","Data":"a441a508256e662aad253db351dca851a1e46457e2b20ce5554b322ae16ffadc"} Dec 09 15:48:29 crc kubenswrapper[4894]: I1209 15:48:29.651530 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-vk5h6" Dec 09 15:48:29 crc kubenswrapper[4894]: I1209 15:48:29.936678 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-vk5h6-config-nr7kn"] Dec 09 15:48:29 crc kubenswrapper[4894]: I1209 15:48:29.944020 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-vk5h6-config-nr7kn"] Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.004695 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-vk5h6-config-2rjz9"] Dec 09 15:48:30 crc kubenswrapper[4894]: E1209 15:48:30.005087 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e8a9251-368a-47a4-872d-da9c17286dd4" containerName="ovn-config" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.005109 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e8a9251-368a-47a4-872d-da9c17286dd4" containerName="ovn-config" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.005322 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e8a9251-368a-47a4-872d-da9c17286dd4" containerName="ovn-config" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.006036 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vk5h6-config-2rjz9" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.010406 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.020888 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vk5h6-config-2rjz9"] Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.080193 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4bc66922-3a12-4dc7-8f13-067b91718ef0-additional-scripts\") pod \"ovn-controller-vk5h6-config-2rjz9\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " pod="openstack/ovn-controller-vk5h6-config-2rjz9" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.080281 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4bc66922-3a12-4dc7-8f13-067b91718ef0-var-run\") pod \"ovn-controller-vk5h6-config-2rjz9\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " pod="openstack/ovn-controller-vk5h6-config-2rjz9" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.080320 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4bc66922-3a12-4dc7-8f13-067b91718ef0-scripts\") pod \"ovn-controller-vk5h6-config-2rjz9\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " pod="openstack/ovn-controller-vk5h6-config-2rjz9" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.080350 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4bc66922-3a12-4dc7-8f13-067b91718ef0-var-run-ovn\") pod \"ovn-controller-vk5h6-config-2rjz9\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " pod="openstack/ovn-controller-vk5h6-config-2rjz9" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.080379 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjvhr\" (UniqueName: \"kubernetes.io/projected/4bc66922-3a12-4dc7-8f13-067b91718ef0-kube-api-access-rjvhr\") pod \"ovn-controller-vk5h6-config-2rjz9\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " pod="openstack/ovn-controller-vk5h6-config-2rjz9" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.080405 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4bc66922-3a12-4dc7-8f13-067b91718ef0-var-log-ovn\") pod \"ovn-controller-vk5h6-config-2rjz9\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " pod="openstack/ovn-controller-vk5h6-config-2rjz9" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.134579 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e8a9251-368a-47a4-872d-da9c17286dd4" path="/var/lib/kubelet/pods/1e8a9251-368a-47a4-872d-da9c17286dd4/volumes" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.182079 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4bc66922-3a12-4dc7-8f13-067b91718ef0-scripts\") pod \"ovn-controller-vk5h6-config-2rjz9\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " pod="openstack/ovn-controller-vk5h6-config-2rjz9" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.182307 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4bc66922-3a12-4dc7-8f13-067b91718ef0-var-run-ovn\") pod \"ovn-controller-vk5h6-config-2rjz9\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " pod="openstack/ovn-controller-vk5h6-config-2rjz9" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.182390 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjvhr\" (UniqueName: \"kubernetes.io/projected/4bc66922-3a12-4dc7-8f13-067b91718ef0-kube-api-access-rjvhr\") pod \"ovn-controller-vk5h6-config-2rjz9\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " pod="openstack/ovn-controller-vk5h6-config-2rjz9" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.182473 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4bc66922-3a12-4dc7-8f13-067b91718ef0-var-log-ovn\") pod \"ovn-controller-vk5h6-config-2rjz9\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " pod="openstack/ovn-controller-vk5h6-config-2rjz9" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.182657 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4bc66922-3a12-4dc7-8f13-067b91718ef0-additional-scripts\") pod \"ovn-controller-vk5h6-config-2rjz9\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " pod="openstack/ovn-controller-vk5h6-config-2rjz9" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.182767 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4bc66922-3a12-4dc7-8f13-067b91718ef0-var-run\") pod \"ovn-controller-vk5h6-config-2rjz9\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " pod="openstack/ovn-controller-vk5h6-config-2rjz9" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.183065 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4bc66922-3a12-4dc7-8f13-067b91718ef0-var-run\") pod \"ovn-controller-vk5h6-config-2rjz9\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " pod="openstack/ovn-controller-vk5h6-config-2rjz9" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.184823 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4bc66922-3a12-4dc7-8f13-067b91718ef0-scripts\") pod \"ovn-controller-vk5h6-config-2rjz9\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " pod="openstack/ovn-controller-vk5h6-config-2rjz9" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.184852 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4bc66922-3a12-4dc7-8f13-067b91718ef0-var-log-ovn\") pod \"ovn-controller-vk5h6-config-2rjz9\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " pod="openstack/ovn-controller-vk5h6-config-2rjz9" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.184898 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4bc66922-3a12-4dc7-8f13-067b91718ef0-var-run-ovn\") pod \"ovn-controller-vk5h6-config-2rjz9\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " pod="openstack/ovn-controller-vk5h6-config-2rjz9" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.185915 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4bc66922-3a12-4dc7-8f13-067b91718ef0-additional-scripts\") pod \"ovn-controller-vk5h6-config-2rjz9\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " pod="openstack/ovn-controller-vk5h6-config-2rjz9" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.201533 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjvhr\" (UniqueName: \"kubernetes.io/projected/4bc66922-3a12-4dc7-8f13-067b91718ef0-kube-api-access-rjvhr\") pod \"ovn-controller-vk5h6-config-2rjz9\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " pod="openstack/ovn-controller-vk5h6-config-2rjz9" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.366185 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vk5h6-config-2rjz9" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.802694 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vk5h6-config-2rjz9"] Dec 09 15:48:30 crc kubenswrapper[4894]: W1209 15:48:30.812799 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4bc66922_3a12_4dc7_8f13_067b91718ef0.slice/crio-fe74fc88d44851d93b4a8e42d76cee33c9daf136eea6886a2fafe8078c99cf6d WatchSource:0}: Error finding container fe74fc88d44851d93b4a8e42d76cee33c9daf136eea6886a2fafe8078c99cf6d: Status 404 returned error can't find the container with id fe74fc88d44851d93b4a8e42d76cee33c9daf136eea6886a2fafe8078c99cf6d Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.906624 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-h74cz" Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.995712 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dwcsb\" (UniqueName: \"kubernetes.io/projected/26cc6eda-fee8-4183-b613-327cc2fad4b6-kube-api-access-dwcsb\") pod \"26cc6eda-fee8-4183-b613-327cc2fad4b6\" (UID: \"26cc6eda-fee8-4183-b613-327cc2fad4b6\") " Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.995986 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26cc6eda-fee8-4183-b613-327cc2fad4b6-combined-ca-bundle\") pod \"26cc6eda-fee8-4183-b613-327cc2fad4b6\" (UID: \"26cc6eda-fee8-4183-b613-327cc2fad4b6\") " Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.996127 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/26cc6eda-fee8-4183-b613-327cc2fad4b6-db-sync-config-data\") pod \"26cc6eda-fee8-4183-b613-327cc2fad4b6\" (UID: \"26cc6eda-fee8-4183-b613-327cc2fad4b6\") " Dec 09 15:48:30 crc kubenswrapper[4894]: I1209 15:48:30.996172 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26cc6eda-fee8-4183-b613-327cc2fad4b6-config-data\") pod \"26cc6eda-fee8-4183-b613-327cc2fad4b6\" (UID: \"26cc6eda-fee8-4183-b613-327cc2fad4b6\") " Dec 09 15:48:31 crc kubenswrapper[4894]: I1209 15:48:31.002119 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26cc6eda-fee8-4183-b613-327cc2fad4b6-kube-api-access-dwcsb" (OuterVolumeSpecName: "kube-api-access-dwcsb") pod "26cc6eda-fee8-4183-b613-327cc2fad4b6" (UID: "26cc6eda-fee8-4183-b613-327cc2fad4b6"). InnerVolumeSpecName "kube-api-access-dwcsb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:48:31 crc kubenswrapper[4894]: I1209 15:48:31.003857 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26cc6eda-fee8-4183-b613-327cc2fad4b6-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "26cc6eda-fee8-4183-b613-327cc2fad4b6" (UID: "26cc6eda-fee8-4183-b613-327cc2fad4b6"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:48:31 crc kubenswrapper[4894]: I1209 15:48:31.030573 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26cc6eda-fee8-4183-b613-327cc2fad4b6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "26cc6eda-fee8-4183-b613-327cc2fad4b6" (UID: "26cc6eda-fee8-4183-b613-327cc2fad4b6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:48:31 crc kubenswrapper[4894]: I1209 15:48:31.043762 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26cc6eda-fee8-4183-b613-327cc2fad4b6-config-data" (OuterVolumeSpecName: "config-data") pod "26cc6eda-fee8-4183-b613-327cc2fad4b6" (UID: "26cc6eda-fee8-4183-b613-327cc2fad4b6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:48:31 crc kubenswrapper[4894]: I1209 15:48:31.097941 4894 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/26cc6eda-fee8-4183-b613-327cc2fad4b6-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:31 crc kubenswrapper[4894]: I1209 15:48:31.097974 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26cc6eda-fee8-4183-b613-327cc2fad4b6-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:31 crc kubenswrapper[4894]: I1209 15:48:31.097985 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dwcsb\" (UniqueName: \"kubernetes.io/projected/26cc6eda-fee8-4183-b613-327cc2fad4b6-kube-api-access-dwcsb\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:31 crc kubenswrapper[4894]: I1209 15:48:31.097997 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/26cc6eda-fee8-4183-b613-327cc2fad4b6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:31 crc kubenswrapper[4894]: I1209 15:48:31.569312 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-h74cz" event={"ID":"26cc6eda-fee8-4183-b613-327cc2fad4b6","Type":"ContainerDied","Data":"cfc934393d82112faa6a48e6f594b5db47169f27af7006a38df31f1aabdbfa08"} Dec 09 15:48:31 crc kubenswrapper[4894]: I1209 15:48:31.569800 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cfc934393d82112faa6a48e6f594b5db47169f27af7006a38df31f1aabdbfa08" Dec 09 15:48:31 crc kubenswrapper[4894]: I1209 15:48:31.569352 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-h74cz" Dec 09 15:48:31 crc kubenswrapper[4894]: I1209 15:48:31.570981 4894 generic.go:334] "Generic (PLEG): container finished" podID="4bc66922-3a12-4dc7-8f13-067b91718ef0" containerID="0b513bac6350ca01ee0afa6381af4c9d11918e749aac91ba1bed011460bb7b59" exitCode=0 Dec 09 15:48:31 crc kubenswrapper[4894]: I1209 15:48:31.571019 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vk5h6-config-2rjz9" event={"ID":"4bc66922-3a12-4dc7-8f13-067b91718ef0","Type":"ContainerDied","Data":"0b513bac6350ca01ee0afa6381af4c9d11918e749aac91ba1bed011460bb7b59"} Dec 09 15:48:31 crc kubenswrapper[4894]: I1209 15:48:31.571041 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vk5h6-config-2rjz9" event={"ID":"4bc66922-3a12-4dc7-8f13-067b91718ef0","Type":"ContainerStarted","Data":"fe74fc88d44851d93b4a8e42d76cee33c9daf136eea6886a2fafe8078c99cf6d"} Dec 09 15:48:31 crc kubenswrapper[4894]: I1209 15:48:31.955931 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5bc56d6f79-xg28g"] Dec 09 15:48:31 crc kubenswrapper[4894]: E1209 15:48:31.956345 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26cc6eda-fee8-4183-b613-327cc2fad4b6" containerName="glance-db-sync" Dec 09 15:48:31 crc kubenswrapper[4894]: I1209 15:48:31.956368 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="26cc6eda-fee8-4183-b613-327cc2fad4b6" containerName="glance-db-sync" Dec 09 15:48:31 crc kubenswrapper[4894]: I1209 15:48:31.956529 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="26cc6eda-fee8-4183-b613-327cc2fad4b6" containerName="glance-db-sync" Dec 09 15:48:31 crc kubenswrapper[4894]: I1209 15:48:31.957655 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.012736 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-config\") pod \"dnsmasq-dns-5bc56d6f79-xg28g\" (UID: \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\") " pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.012835 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-dns-svc\") pod \"dnsmasq-dns-5bc56d6f79-xg28g\" (UID: \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\") " pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.012901 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-ovsdbserver-nb\") pod \"dnsmasq-dns-5bc56d6f79-xg28g\" (UID: \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\") " pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.012937 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-ovsdbserver-sb\") pod \"dnsmasq-dns-5bc56d6f79-xg28g\" (UID: \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\") " pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.012970 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p4m8h\" (UniqueName: \"kubernetes.io/projected/5c6eec80-a266-4965-9b10-c5e7b9dee04d-kube-api-access-p4m8h\") pod \"dnsmasq-dns-5bc56d6f79-xg28g\" (UID: \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\") " pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.017027 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bc56d6f79-xg28g"] Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.114340 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-config\") pod \"dnsmasq-dns-5bc56d6f79-xg28g\" (UID: \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\") " pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.114468 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-dns-svc\") pod \"dnsmasq-dns-5bc56d6f79-xg28g\" (UID: \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\") " pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.114603 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-ovsdbserver-nb\") pod \"dnsmasq-dns-5bc56d6f79-xg28g\" (UID: \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\") " pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.114669 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-ovsdbserver-sb\") pod \"dnsmasq-dns-5bc56d6f79-xg28g\" (UID: \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\") " pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.114712 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p4m8h\" (UniqueName: \"kubernetes.io/projected/5c6eec80-a266-4965-9b10-c5e7b9dee04d-kube-api-access-p4m8h\") pod \"dnsmasq-dns-5bc56d6f79-xg28g\" (UID: \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\") " pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.115585 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-ovsdbserver-nb\") pod \"dnsmasq-dns-5bc56d6f79-xg28g\" (UID: \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\") " pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.115658 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-config\") pod \"dnsmasq-dns-5bc56d6f79-xg28g\" (UID: \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\") " pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.115727 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-dns-svc\") pod \"dnsmasq-dns-5bc56d6f79-xg28g\" (UID: \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\") " pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.115855 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-ovsdbserver-sb\") pod \"dnsmasq-dns-5bc56d6f79-xg28g\" (UID: \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\") " pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.137381 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p4m8h\" (UniqueName: \"kubernetes.io/projected/5c6eec80-a266-4965-9b10-c5e7b9dee04d-kube-api-access-p4m8h\") pod \"dnsmasq-dns-5bc56d6f79-xg28g\" (UID: \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\") " pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.273600 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.709970 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5bc56d6f79-xg28g"] Dec 09 15:48:32 crc kubenswrapper[4894]: W1209 15:48:32.740880 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c6eec80_a266_4965_9b10_c5e7b9dee04d.slice/crio-75129d3aa213adf22467d039f507f3b3892efcb2947e7979cdce3c4efe514e8b WatchSource:0}: Error finding container 75129d3aa213adf22467d039f507f3b3892efcb2947e7979cdce3c4efe514e8b: Status 404 returned error can't find the container with id 75129d3aa213adf22467d039f507f3b3892efcb2947e7979cdce3c4efe514e8b Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.891924 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vk5h6-config-2rjz9" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.926761 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4bc66922-3a12-4dc7-8f13-067b91718ef0-var-run\") pod \"4bc66922-3a12-4dc7-8f13-067b91718ef0\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.926881 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4bc66922-3a12-4dc7-8f13-067b91718ef0-var-run" (OuterVolumeSpecName: "var-run") pod "4bc66922-3a12-4dc7-8f13-067b91718ef0" (UID: "4bc66922-3a12-4dc7-8f13-067b91718ef0"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.926904 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4bc66922-3a12-4dc7-8f13-067b91718ef0-additional-scripts\") pod \"4bc66922-3a12-4dc7-8f13-067b91718ef0\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.927041 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4bc66922-3a12-4dc7-8f13-067b91718ef0-scripts\") pod \"4bc66922-3a12-4dc7-8f13-067b91718ef0\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.927083 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4bc66922-3a12-4dc7-8f13-067b91718ef0-var-run-ovn\") pod \"4bc66922-3a12-4dc7-8f13-067b91718ef0\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.927132 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4bc66922-3a12-4dc7-8f13-067b91718ef0-var-log-ovn\") pod \"4bc66922-3a12-4dc7-8f13-067b91718ef0\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.927353 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rjvhr\" (UniqueName: \"kubernetes.io/projected/4bc66922-3a12-4dc7-8f13-067b91718ef0-kube-api-access-rjvhr\") pod \"4bc66922-3a12-4dc7-8f13-067b91718ef0\" (UID: \"4bc66922-3a12-4dc7-8f13-067b91718ef0\") " Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.927380 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4bc66922-3a12-4dc7-8f13-067b91718ef0-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "4bc66922-3a12-4dc7-8f13-067b91718ef0" (UID: "4bc66922-3a12-4dc7-8f13-067b91718ef0"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.927425 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4bc66922-3a12-4dc7-8f13-067b91718ef0-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "4bc66922-3a12-4dc7-8f13-067b91718ef0" (UID: "4bc66922-3a12-4dc7-8f13-067b91718ef0"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.928057 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bc66922-3a12-4dc7-8f13-067b91718ef0-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "4bc66922-3a12-4dc7-8f13-067b91718ef0" (UID: "4bc66922-3a12-4dc7-8f13-067b91718ef0"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.928074 4894 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4bc66922-3a12-4dc7-8f13-067b91718ef0-var-run\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.928098 4894 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/4bc66922-3a12-4dc7-8f13-067b91718ef0-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.928113 4894 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/4bc66922-3a12-4dc7-8f13-067b91718ef0-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.928610 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bc66922-3a12-4dc7-8f13-067b91718ef0-scripts" (OuterVolumeSpecName: "scripts") pod "4bc66922-3a12-4dc7-8f13-067b91718ef0" (UID: "4bc66922-3a12-4dc7-8f13-067b91718ef0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:32 crc kubenswrapper[4894]: I1209 15:48:32.933095 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bc66922-3a12-4dc7-8f13-067b91718ef0-kube-api-access-rjvhr" (OuterVolumeSpecName: "kube-api-access-rjvhr") pod "4bc66922-3a12-4dc7-8f13-067b91718ef0" (UID: "4bc66922-3a12-4dc7-8f13-067b91718ef0"). InnerVolumeSpecName "kube-api-access-rjvhr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:48:33 crc kubenswrapper[4894]: I1209 15:48:33.029605 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4bc66922-3a12-4dc7-8f13-067b91718ef0-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:33 crc kubenswrapper[4894]: I1209 15:48:33.029666 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rjvhr\" (UniqueName: \"kubernetes.io/projected/4bc66922-3a12-4dc7-8f13-067b91718ef0-kube-api-access-rjvhr\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:33 crc kubenswrapper[4894]: I1209 15:48:33.029682 4894 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/4bc66922-3a12-4dc7-8f13-067b91718ef0-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:33 crc kubenswrapper[4894]: I1209 15:48:33.589951 4894 generic.go:334] "Generic (PLEG): container finished" podID="5c6eec80-a266-4965-9b10-c5e7b9dee04d" containerID="eab5597ec0b22d524b0f4e7b62f0df5bc6c0595217f9c4687712ca2d09758473" exitCode=0 Dec 09 15:48:33 crc kubenswrapper[4894]: I1209 15:48:33.590062 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" event={"ID":"5c6eec80-a266-4965-9b10-c5e7b9dee04d","Type":"ContainerDied","Data":"eab5597ec0b22d524b0f4e7b62f0df5bc6c0595217f9c4687712ca2d09758473"} Dec 09 15:48:33 crc kubenswrapper[4894]: I1209 15:48:33.590138 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" event={"ID":"5c6eec80-a266-4965-9b10-c5e7b9dee04d","Type":"ContainerStarted","Data":"75129d3aa213adf22467d039f507f3b3892efcb2947e7979cdce3c4efe514e8b"} Dec 09 15:48:33 crc kubenswrapper[4894]: I1209 15:48:33.591982 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vk5h6-config-2rjz9" event={"ID":"4bc66922-3a12-4dc7-8f13-067b91718ef0","Type":"ContainerDied","Data":"fe74fc88d44851d93b4a8e42d76cee33c9daf136eea6886a2fafe8078c99cf6d"} Dec 09 15:48:33 crc kubenswrapper[4894]: I1209 15:48:33.592029 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe74fc88d44851d93b4a8e42d76cee33c9daf136eea6886a2fafe8078c99cf6d" Dec 09 15:48:33 crc kubenswrapper[4894]: I1209 15:48:33.592037 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vk5h6-config-2rjz9" Dec 09 15:48:33 crc kubenswrapper[4894]: I1209 15:48:33.993502 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-vk5h6-config-2rjz9"] Dec 09 15:48:33 crc kubenswrapper[4894]: I1209 15:48:33.999929 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-vk5h6-config-2rjz9"] Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.033893 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-vk5h6-config-hlk99"] Dec 09 15:48:34 crc kubenswrapper[4894]: E1209 15:48:34.034309 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bc66922-3a12-4dc7-8f13-067b91718ef0" containerName="ovn-config" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.034324 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bc66922-3a12-4dc7-8f13-067b91718ef0" containerName="ovn-config" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.034517 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bc66922-3a12-4dc7-8f13-067b91718ef0" containerName="ovn-config" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.035224 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vk5h6-config-hlk99" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.037403 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.046332 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vk5h6-config-hlk99"] Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.117868 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bc66922-3a12-4dc7-8f13-067b91718ef0" path="/var/lib/kubelet/pods/4bc66922-3a12-4dc7-8f13-067b91718ef0/volumes" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.148170 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/aa46f14a-27b9-4d99-aeed-63dc95f536ee-var-run-ovn\") pod \"ovn-controller-vk5h6-config-hlk99\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " pod="openstack/ovn-controller-vk5h6-config-hlk99" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.148238 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/aa46f14a-27b9-4d99-aeed-63dc95f536ee-additional-scripts\") pod \"ovn-controller-vk5h6-config-hlk99\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " pod="openstack/ovn-controller-vk5h6-config-hlk99" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.148286 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/aa46f14a-27b9-4d99-aeed-63dc95f536ee-var-log-ovn\") pod \"ovn-controller-vk5h6-config-hlk99\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " pod="openstack/ovn-controller-vk5h6-config-hlk99" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.148316 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swwz5\" (UniqueName: \"kubernetes.io/projected/aa46f14a-27b9-4d99-aeed-63dc95f536ee-kube-api-access-swwz5\") pod \"ovn-controller-vk5h6-config-hlk99\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " pod="openstack/ovn-controller-vk5h6-config-hlk99" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.148385 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aa46f14a-27b9-4d99-aeed-63dc95f536ee-var-run\") pod \"ovn-controller-vk5h6-config-hlk99\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " pod="openstack/ovn-controller-vk5h6-config-hlk99" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.148460 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa46f14a-27b9-4d99-aeed-63dc95f536ee-scripts\") pod \"ovn-controller-vk5h6-config-hlk99\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " pod="openstack/ovn-controller-vk5h6-config-hlk99" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.250322 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/aa46f14a-27b9-4d99-aeed-63dc95f536ee-additional-scripts\") pod \"ovn-controller-vk5h6-config-hlk99\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " pod="openstack/ovn-controller-vk5h6-config-hlk99" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.250372 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/aa46f14a-27b9-4d99-aeed-63dc95f536ee-var-log-ovn\") pod \"ovn-controller-vk5h6-config-hlk99\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " pod="openstack/ovn-controller-vk5h6-config-hlk99" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.250397 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swwz5\" (UniqueName: \"kubernetes.io/projected/aa46f14a-27b9-4d99-aeed-63dc95f536ee-kube-api-access-swwz5\") pod \"ovn-controller-vk5h6-config-hlk99\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " pod="openstack/ovn-controller-vk5h6-config-hlk99" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.250454 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aa46f14a-27b9-4d99-aeed-63dc95f536ee-var-run\") pod \"ovn-controller-vk5h6-config-hlk99\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " pod="openstack/ovn-controller-vk5h6-config-hlk99" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.250484 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa46f14a-27b9-4d99-aeed-63dc95f536ee-scripts\") pod \"ovn-controller-vk5h6-config-hlk99\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " pod="openstack/ovn-controller-vk5h6-config-hlk99" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.250533 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/aa46f14a-27b9-4d99-aeed-63dc95f536ee-var-run-ovn\") pod \"ovn-controller-vk5h6-config-hlk99\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " pod="openstack/ovn-controller-vk5h6-config-hlk99" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.250817 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/aa46f14a-27b9-4d99-aeed-63dc95f536ee-var-run-ovn\") pod \"ovn-controller-vk5h6-config-hlk99\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " pod="openstack/ovn-controller-vk5h6-config-hlk99" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.250835 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aa46f14a-27b9-4d99-aeed-63dc95f536ee-var-run\") pod \"ovn-controller-vk5h6-config-hlk99\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " pod="openstack/ovn-controller-vk5h6-config-hlk99" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.250855 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/aa46f14a-27b9-4d99-aeed-63dc95f536ee-var-log-ovn\") pod \"ovn-controller-vk5h6-config-hlk99\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " pod="openstack/ovn-controller-vk5h6-config-hlk99" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.251254 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/aa46f14a-27b9-4d99-aeed-63dc95f536ee-additional-scripts\") pod \"ovn-controller-vk5h6-config-hlk99\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " pod="openstack/ovn-controller-vk5h6-config-hlk99" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.252842 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa46f14a-27b9-4d99-aeed-63dc95f536ee-scripts\") pod \"ovn-controller-vk5h6-config-hlk99\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " pod="openstack/ovn-controller-vk5h6-config-hlk99" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.269547 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swwz5\" (UniqueName: \"kubernetes.io/projected/aa46f14a-27b9-4d99-aeed-63dc95f536ee-kube-api-access-swwz5\") pod \"ovn-controller-vk5h6-config-hlk99\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " pod="openstack/ovn-controller-vk5h6-config-hlk99" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.356728 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vk5h6-config-hlk99" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.607789 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" event={"ID":"5c6eec80-a266-4965-9b10-c5e7b9dee04d","Type":"ContainerStarted","Data":"bd664398729728de264c1e429a821521b44bbc925e4855a7292f30c11396b30e"} Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.608886 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.628490 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" podStartSLOduration=3.628467029 podStartE2EDuration="3.628467029s" podCreationTimestamp="2025-12-09 15:48:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:48:34.628136381 +0000 UTC m=+1008.947347050" watchObservedRunningTime="2025-12-09 15:48:34.628467029 +0000 UTC m=+1008.947677698" Dec 09 15:48:34 crc kubenswrapper[4894]: I1209 15:48:34.834477 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vk5h6-config-hlk99"] Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.075845 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.372185 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-pfn9g"] Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.373569 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-pfn9g" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.394210 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-pfn9g"] Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.399854 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.502577 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fqmhp\" (UniqueName: \"kubernetes.io/projected/a754121b-d60b-4107-ac9c-15cc6db44d93-kube-api-access-fqmhp\") pod \"cinder-db-create-pfn9g\" (UID: \"a754121b-d60b-4107-ac9c-15cc6db44d93\") " pod="openstack/cinder-db-create-pfn9g" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.502982 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a754121b-d60b-4107-ac9c-15cc6db44d93-operator-scripts\") pod \"cinder-db-create-pfn9g\" (UID: \"a754121b-d60b-4107-ac9c-15cc6db44d93\") " pod="openstack/cinder-db-create-pfn9g" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.509357 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-wpqxn"] Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.510567 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-wpqxn" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.531846 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-a57d-account-create-update-8xxg4"] Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.532850 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-a57d-account-create-update-8xxg4" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.534839 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.555392 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-a57d-account-create-update-8xxg4"] Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.569261 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-wpqxn"] Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.595726 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-fc79-account-create-update-ts2zw"] Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.597971 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-fc79-account-create-update-ts2zw" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.600386 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.604618 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fqmhp\" (UniqueName: \"kubernetes.io/projected/a754121b-d60b-4107-ac9c-15cc6db44d93-kube-api-access-fqmhp\") pod \"cinder-db-create-pfn9g\" (UID: \"a754121b-d60b-4107-ac9c-15cc6db44d93\") " pod="openstack/cinder-db-create-pfn9g" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.604839 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a754121b-d60b-4107-ac9c-15cc6db44d93-operator-scripts\") pod \"cinder-db-create-pfn9g\" (UID: \"a754121b-d60b-4107-ac9c-15cc6db44d93\") " pod="openstack/cinder-db-create-pfn9g" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.606367 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a754121b-d60b-4107-ac9c-15cc6db44d93-operator-scripts\") pod \"cinder-db-create-pfn9g\" (UID: \"a754121b-d60b-4107-ac9c-15cc6db44d93\") " pod="openstack/cinder-db-create-pfn9g" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.615521 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-fc79-account-create-update-ts2zw"] Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.618906 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vk5h6-config-hlk99" event={"ID":"aa46f14a-27b9-4d99-aeed-63dc95f536ee","Type":"ContainerStarted","Data":"7031c978f263c0eca3e326f932796ae71acdc505f759c19d24b9ce6a759aa6fb"} Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.618950 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vk5h6-config-hlk99" event={"ID":"aa46f14a-27b9-4d99-aeed-63dc95f536ee","Type":"ContainerStarted","Data":"54ec4e5ece5839717f6d174f6298505c1a9eb9a44ef1e416139685121b22d4c2"} Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.639559 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fqmhp\" (UniqueName: \"kubernetes.io/projected/a754121b-d60b-4107-ac9c-15cc6db44d93-kube-api-access-fqmhp\") pod \"cinder-db-create-pfn9g\" (UID: \"a754121b-d60b-4107-ac9c-15cc6db44d93\") " pod="openstack/cinder-db-create-pfn9g" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.653455 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-vk5h6-config-hlk99" podStartSLOduration=1.6534388880000002 podStartE2EDuration="1.653438888s" podCreationTimestamp="2025-12-09 15:48:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:48:35.651046775 +0000 UTC m=+1009.970257444" watchObservedRunningTime="2025-12-09 15:48:35.653438888 +0000 UTC m=+1009.972649557" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.691588 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-pfn9g" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.693706 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-npqvf"] Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.695245 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-npqvf" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.698917 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.699342 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-g8fmh" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.699866 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.700063 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.706817 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w2vb4\" (UniqueName: \"kubernetes.io/projected/946604b5-e779-4579-a593-f46a7c4e3e5e-kube-api-access-w2vb4\") pod \"barbican-fc79-account-create-update-ts2zw\" (UID: \"946604b5-e779-4579-a593-f46a7c4e3e5e\") " pod="openstack/barbican-fc79-account-create-update-ts2zw" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.707020 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggl5x\" (UniqueName: \"kubernetes.io/projected/b191efc2-e3bf-49e8-83d9-7b2e55264863-kube-api-access-ggl5x\") pod \"barbican-db-create-wpqxn\" (UID: \"b191efc2-e3bf-49e8-83d9-7b2e55264863\") " pod="openstack/barbican-db-create-wpqxn" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.707053 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09b8ca6e-7521-4e7f-8272-504d2d7d66d7-operator-scripts\") pod \"cinder-a57d-account-create-update-8xxg4\" (UID: \"09b8ca6e-7521-4e7f-8272-504d2d7d66d7\") " pod="openstack/cinder-a57d-account-create-update-8xxg4" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.707080 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b191efc2-e3bf-49e8-83d9-7b2e55264863-operator-scripts\") pod \"barbican-db-create-wpqxn\" (UID: \"b191efc2-e3bf-49e8-83d9-7b2e55264863\") " pod="openstack/barbican-db-create-wpqxn" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.707119 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/946604b5-e779-4579-a593-f46a7c4e3e5e-operator-scripts\") pod \"barbican-fc79-account-create-update-ts2zw\" (UID: \"946604b5-e779-4579-a593-f46a7c4e3e5e\") " pod="openstack/barbican-fc79-account-create-update-ts2zw" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.707146 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrhc6\" (UniqueName: \"kubernetes.io/projected/09b8ca6e-7521-4e7f-8272-504d2d7d66d7-kube-api-access-xrhc6\") pod \"cinder-a57d-account-create-update-8xxg4\" (UID: \"09b8ca6e-7521-4e7f-8272-504d2d7d66d7\") " pod="openstack/cinder-a57d-account-create-update-8xxg4" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.711131 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-npqvf"] Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.777440 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-v6zx2"] Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.779799 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-v6zx2" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.799703 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-v6zx2"] Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.808618 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87072774-8453-49d2-a355-3388179d06dc-config-data\") pod \"keystone-db-sync-npqvf\" (UID: \"87072774-8453-49d2-a355-3388179d06dc\") " pod="openstack/keystone-db-sync-npqvf" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.808714 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87072774-8453-49d2-a355-3388179d06dc-combined-ca-bundle\") pod \"keystone-db-sync-npqvf\" (UID: \"87072774-8453-49d2-a355-3388179d06dc\") " pod="openstack/keystone-db-sync-npqvf" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.808764 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggl5x\" (UniqueName: \"kubernetes.io/projected/b191efc2-e3bf-49e8-83d9-7b2e55264863-kube-api-access-ggl5x\") pod \"barbican-db-create-wpqxn\" (UID: \"b191efc2-e3bf-49e8-83d9-7b2e55264863\") " pod="openstack/barbican-db-create-wpqxn" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.808788 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09b8ca6e-7521-4e7f-8272-504d2d7d66d7-operator-scripts\") pod \"cinder-a57d-account-create-update-8xxg4\" (UID: \"09b8ca6e-7521-4e7f-8272-504d2d7d66d7\") " pod="openstack/cinder-a57d-account-create-update-8xxg4" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.808817 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b191efc2-e3bf-49e8-83d9-7b2e55264863-operator-scripts\") pod \"barbican-db-create-wpqxn\" (UID: \"b191efc2-e3bf-49e8-83d9-7b2e55264863\") " pod="openstack/barbican-db-create-wpqxn" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.808851 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/946604b5-e779-4579-a593-f46a7c4e3e5e-operator-scripts\") pod \"barbican-fc79-account-create-update-ts2zw\" (UID: \"946604b5-e779-4579-a593-f46a7c4e3e5e\") " pod="openstack/barbican-fc79-account-create-update-ts2zw" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.808877 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8xxzk\" (UniqueName: \"kubernetes.io/projected/87072774-8453-49d2-a355-3388179d06dc-kube-api-access-8xxzk\") pod \"keystone-db-sync-npqvf\" (UID: \"87072774-8453-49d2-a355-3388179d06dc\") " pod="openstack/keystone-db-sync-npqvf" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.808909 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrhc6\" (UniqueName: \"kubernetes.io/projected/09b8ca6e-7521-4e7f-8272-504d2d7d66d7-kube-api-access-xrhc6\") pod \"cinder-a57d-account-create-update-8xxg4\" (UID: \"09b8ca6e-7521-4e7f-8272-504d2d7d66d7\") " pod="openstack/cinder-a57d-account-create-update-8xxg4" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.808984 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w2vb4\" (UniqueName: \"kubernetes.io/projected/946604b5-e779-4579-a593-f46a7c4e3e5e-kube-api-access-w2vb4\") pod \"barbican-fc79-account-create-update-ts2zw\" (UID: \"946604b5-e779-4579-a593-f46a7c4e3e5e\") " pod="openstack/barbican-fc79-account-create-update-ts2zw" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.810373 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09b8ca6e-7521-4e7f-8272-504d2d7d66d7-operator-scripts\") pod \"cinder-a57d-account-create-update-8xxg4\" (UID: \"09b8ca6e-7521-4e7f-8272-504d2d7d66d7\") " pod="openstack/cinder-a57d-account-create-update-8xxg4" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.810374 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b191efc2-e3bf-49e8-83d9-7b2e55264863-operator-scripts\") pod \"barbican-db-create-wpqxn\" (UID: \"b191efc2-e3bf-49e8-83d9-7b2e55264863\") " pod="openstack/barbican-db-create-wpqxn" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.810910 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/946604b5-e779-4579-a593-f46a7c4e3e5e-operator-scripts\") pod \"barbican-fc79-account-create-update-ts2zw\" (UID: \"946604b5-e779-4579-a593-f46a7c4e3e5e\") " pod="openstack/barbican-fc79-account-create-update-ts2zw" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.831329 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggl5x\" (UniqueName: \"kubernetes.io/projected/b191efc2-e3bf-49e8-83d9-7b2e55264863-kube-api-access-ggl5x\") pod \"barbican-db-create-wpqxn\" (UID: \"b191efc2-e3bf-49e8-83d9-7b2e55264863\") " pod="openstack/barbican-db-create-wpqxn" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.837802 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w2vb4\" (UniqueName: \"kubernetes.io/projected/946604b5-e779-4579-a593-f46a7c4e3e5e-kube-api-access-w2vb4\") pod \"barbican-fc79-account-create-update-ts2zw\" (UID: \"946604b5-e779-4579-a593-f46a7c4e3e5e\") " pod="openstack/barbican-fc79-account-create-update-ts2zw" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.838232 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrhc6\" (UniqueName: \"kubernetes.io/projected/09b8ca6e-7521-4e7f-8272-504d2d7d66d7-kube-api-access-xrhc6\") pod \"cinder-a57d-account-create-update-8xxg4\" (UID: \"09b8ca6e-7521-4e7f-8272-504d2d7d66d7\") " pod="openstack/cinder-a57d-account-create-update-8xxg4" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.850302 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-a57d-account-create-update-8xxg4" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.889349 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5da2-account-create-update-wthfq"] Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.908258 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5da2-account-create-update-wthfq"] Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.908773 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5da2-account-create-update-wthfq" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.916319 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-fc79-account-create-update-ts2zw" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.916551 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8xxzk\" (UniqueName: \"kubernetes.io/projected/87072774-8453-49d2-a355-3388179d06dc-kube-api-access-8xxzk\") pod \"keystone-db-sync-npqvf\" (UID: \"87072774-8453-49d2-a355-3388179d06dc\") " pod="openstack/keystone-db-sync-npqvf" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.917804 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dl7xf\" (UniqueName: \"kubernetes.io/projected/48cc59c5-cf8b-455b-8f27-beb7398510b3-kube-api-access-dl7xf\") pod \"neutron-db-create-v6zx2\" (UID: \"48cc59c5-cf8b-455b-8f27-beb7398510b3\") " pod="openstack/neutron-db-create-v6zx2" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.917943 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87072774-8453-49d2-a355-3388179d06dc-config-data\") pod \"keystone-db-sync-npqvf\" (UID: \"87072774-8453-49d2-a355-3388179d06dc\") " pod="openstack/keystone-db-sync-npqvf" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.918050 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87072774-8453-49d2-a355-3388179d06dc-combined-ca-bundle\") pod \"keystone-db-sync-npqvf\" (UID: \"87072774-8453-49d2-a355-3388179d06dc\") " pod="openstack/keystone-db-sync-npqvf" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.918186 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48cc59c5-cf8b-455b-8f27-beb7398510b3-operator-scripts\") pod \"neutron-db-create-v6zx2\" (UID: \"48cc59c5-cf8b-455b-8f27-beb7398510b3\") " pod="openstack/neutron-db-create-v6zx2" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.923168 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87072774-8453-49d2-a355-3388179d06dc-config-data\") pod \"keystone-db-sync-npqvf\" (UID: \"87072774-8453-49d2-a355-3388179d06dc\") " pod="openstack/keystone-db-sync-npqvf" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.927350 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87072774-8453-49d2-a355-3388179d06dc-combined-ca-bundle\") pod \"keystone-db-sync-npqvf\" (UID: \"87072774-8453-49d2-a355-3388179d06dc\") " pod="openstack/keystone-db-sync-npqvf" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.929922 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Dec 09 15:48:35 crc kubenswrapper[4894]: I1209 15:48:35.944899 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8xxzk\" (UniqueName: \"kubernetes.io/projected/87072774-8453-49d2-a355-3388179d06dc-kube-api-access-8xxzk\") pod \"keystone-db-sync-npqvf\" (UID: \"87072774-8453-49d2-a355-3388179d06dc\") " pod="openstack/keystone-db-sync-npqvf" Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.020127 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dl7xf\" (UniqueName: \"kubernetes.io/projected/48cc59c5-cf8b-455b-8f27-beb7398510b3-kube-api-access-dl7xf\") pod \"neutron-db-create-v6zx2\" (UID: \"48cc59c5-cf8b-455b-8f27-beb7398510b3\") " pod="openstack/neutron-db-create-v6zx2" Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.020242 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48cc59c5-cf8b-455b-8f27-beb7398510b3-operator-scripts\") pod \"neutron-db-create-v6zx2\" (UID: \"48cc59c5-cf8b-455b-8f27-beb7398510b3\") " pod="openstack/neutron-db-create-v6zx2" Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.020277 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81e9c778-7d46-47f9-8d63-15662adc4aa3-operator-scripts\") pod \"neutron-5da2-account-create-update-wthfq\" (UID: \"81e9c778-7d46-47f9-8d63-15662adc4aa3\") " pod="openstack/neutron-5da2-account-create-update-wthfq" Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.020325 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gp2d9\" (UniqueName: \"kubernetes.io/projected/81e9c778-7d46-47f9-8d63-15662adc4aa3-kube-api-access-gp2d9\") pod \"neutron-5da2-account-create-update-wthfq\" (UID: \"81e9c778-7d46-47f9-8d63-15662adc4aa3\") " pod="openstack/neutron-5da2-account-create-update-wthfq" Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.021803 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48cc59c5-cf8b-455b-8f27-beb7398510b3-operator-scripts\") pod \"neutron-db-create-v6zx2\" (UID: \"48cc59c5-cf8b-455b-8f27-beb7398510b3\") " pod="openstack/neutron-db-create-v6zx2" Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.026789 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-npqvf" Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.037300 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dl7xf\" (UniqueName: \"kubernetes.io/projected/48cc59c5-cf8b-455b-8f27-beb7398510b3-kube-api-access-dl7xf\") pod \"neutron-db-create-v6zx2\" (UID: \"48cc59c5-cf8b-455b-8f27-beb7398510b3\") " pod="openstack/neutron-db-create-v6zx2" Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.121529 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81e9c778-7d46-47f9-8d63-15662adc4aa3-operator-scripts\") pod \"neutron-5da2-account-create-update-wthfq\" (UID: \"81e9c778-7d46-47f9-8d63-15662adc4aa3\") " pod="openstack/neutron-5da2-account-create-update-wthfq" Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.121923 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gp2d9\" (UniqueName: \"kubernetes.io/projected/81e9c778-7d46-47f9-8d63-15662adc4aa3-kube-api-access-gp2d9\") pod \"neutron-5da2-account-create-update-wthfq\" (UID: \"81e9c778-7d46-47f9-8d63-15662adc4aa3\") " pod="openstack/neutron-5da2-account-create-update-wthfq" Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.122528 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81e9c778-7d46-47f9-8d63-15662adc4aa3-operator-scripts\") pod \"neutron-5da2-account-create-update-wthfq\" (UID: \"81e9c778-7d46-47f9-8d63-15662adc4aa3\") " pod="openstack/neutron-5da2-account-create-update-wthfq" Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.131276 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-wpqxn" Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.139056 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-pfn9g"] Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.143208 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gp2d9\" (UniqueName: \"kubernetes.io/projected/81e9c778-7d46-47f9-8d63-15662adc4aa3-kube-api-access-gp2d9\") pod \"neutron-5da2-account-create-update-wthfq\" (UID: \"81e9c778-7d46-47f9-8d63-15662adc4aa3\") " pod="openstack/neutron-5da2-account-create-update-wthfq" Dec 09 15:48:36 crc kubenswrapper[4894]: W1209 15:48:36.165772 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda754121b_d60b_4107_ac9c_15cc6db44d93.slice/crio-7d41feca8c72d8a58b64feec1609138aa5df13c8659d94cf6fafd7fc61f9c7d5 WatchSource:0}: Error finding container 7d41feca8c72d8a58b64feec1609138aa5df13c8659d94cf6fafd7fc61f9c7d5: Status 404 returned error can't find the container with id 7d41feca8c72d8a58b64feec1609138aa5df13c8659d94cf6fafd7fc61f9c7d5 Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.231859 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-v6zx2" Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.254600 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5da2-account-create-update-wthfq" Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.443242 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-fc79-account-create-update-ts2zw"] Dec 09 15:48:36 crc kubenswrapper[4894]: W1209 15:48:36.457907 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09b8ca6e_7521_4e7f_8272_504d2d7d66d7.slice/crio-083214a4b8bca511aade9bc4f97cc155e08a9e4e7230cd3072c92e617733d305 WatchSource:0}: Error finding container 083214a4b8bca511aade9bc4f97cc155e08a9e4e7230cd3072c92e617733d305: Status 404 returned error can't find the container with id 083214a4b8bca511aade9bc4f97cc155e08a9e4e7230cd3072c92e617733d305 Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.473080 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-a57d-account-create-update-8xxg4"] Dec 09 15:48:36 crc kubenswrapper[4894]: W1209 15:48:36.492166 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod946604b5_e779_4579_a593_f46a7c4e3e5e.slice/crio-70ba905c2a3749be73032e64b0960018e62eba975c552ab97f6496c9bdaad786 WatchSource:0}: Error finding container 70ba905c2a3749be73032e64b0960018e62eba975c552ab97f6496c9bdaad786: Status 404 returned error can't find the container with id 70ba905c2a3749be73032e64b0960018e62eba975c552ab97f6496c9bdaad786 Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.492571 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-wpqxn"] Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.536254 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-npqvf"] Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.634798 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-fc79-account-create-update-ts2zw" event={"ID":"946604b5-e779-4579-a593-f46a7c4e3e5e","Type":"ContainerStarted","Data":"70ba905c2a3749be73032e64b0960018e62eba975c552ab97f6496c9bdaad786"} Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.641189 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-a57d-account-create-update-8xxg4" event={"ID":"09b8ca6e-7521-4e7f-8272-504d2d7d66d7","Type":"ContainerStarted","Data":"c690018c1d490040c1720247120a6a74843ecabd9c34e41fb997b014f699bf4c"} Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.641239 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-a57d-account-create-update-8xxg4" event={"ID":"09b8ca6e-7521-4e7f-8272-504d2d7d66d7","Type":"ContainerStarted","Data":"083214a4b8bca511aade9bc4f97cc155e08a9e4e7230cd3072c92e617733d305"} Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.643329 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-wpqxn" event={"ID":"b191efc2-e3bf-49e8-83d9-7b2e55264863","Type":"ContainerStarted","Data":"1a961f4663ff5fd098c4610a9ffff4d556636698ccbc761c2ffde5eabdbe33db"} Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.649287 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vk5h6-config-hlk99" event={"ID":"aa46f14a-27b9-4d99-aeed-63dc95f536ee","Type":"ContainerDied","Data":"7031c978f263c0eca3e326f932796ae71acdc505f759c19d24b9ce6a759aa6fb"} Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.649306 4894 generic.go:334] "Generic (PLEG): container finished" podID="aa46f14a-27b9-4d99-aeed-63dc95f536ee" containerID="7031c978f263c0eca3e326f932796ae71acdc505f759c19d24b9ce6a759aa6fb" exitCode=0 Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.652088 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-pfn9g" event={"ID":"a754121b-d60b-4107-ac9c-15cc6db44d93","Type":"ContainerStarted","Data":"723467964f41a9aa68ad1cbed609234ca51ea9e01d04854c81afa5b441925f83"} Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.652114 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-pfn9g" event={"ID":"a754121b-d60b-4107-ac9c-15cc6db44d93","Type":"ContainerStarted","Data":"7d41feca8c72d8a58b64feec1609138aa5df13c8659d94cf6fafd7fc61f9c7d5"} Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.656269 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-npqvf" event={"ID":"87072774-8453-49d2-a355-3388179d06dc","Type":"ContainerStarted","Data":"3437db2c3e7023d1f04456ab6d82e8c3ba39ff11ccf2cd533ff549d5cdc4a175"} Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.676912 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-a57d-account-create-update-8xxg4" podStartSLOduration=1.676888076 podStartE2EDuration="1.676888076s" podCreationTimestamp="2025-12-09 15:48:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:48:36.660303513 +0000 UTC m=+1010.979514182" watchObservedRunningTime="2025-12-09 15:48:36.676888076 +0000 UTC m=+1010.996098745" Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.719361 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-pfn9g" podStartSLOduration=1.719343996 podStartE2EDuration="1.719343996s" podCreationTimestamp="2025-12-09 15:48:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:48:36.716948033 +0000 UTC m=+1011.036158722" watchObservedRunningTime="2025-12-09 15:48:36.719343996 +0000 UTC m=+1011.038554665" Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.782422 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-v6zx2"] Dec 09 15:48:36 crc kubenswrapper[4894]: W1209 15:48:36.790625 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod48cc59c5_cf8b_455b_8f27_beb7398510b3.slice/crio-c71acd0df855104137898aea030c6a8a34375e3c25b41bc9997c7f33519dc302 WatchSource:0}: Error finding container c71acd0df855104137898aea030c6a8a34375e3c25b41bc9997c7f33519dc302: Status 404 returned error can't find the container with id c71acd0df855104137898aea030c6a8a34375e3c25b41bc9997c7f33519dc302 Dec 09 15:48:36 crc kubenswrapper[4894]: I1209 15:48:36.869144 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5da2-account-create-update-wthfq"] Dec 09 15:48:36 crc kubenswrapper[4894]: W1209 15:48:36.888469 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod81e9c778_7d46_47f9_8d63_15662adc4aa3.slice/crio-8079eaecc6f5238a9a35b5f70c66f523d1598acaec34188c2dbb80b4ffcde39c WatchSource:0}: Error finding container 8079eaecc6f5238a9a35b5f70c66f523d1598acaec34188c2dbb80b4ffcde39c: Status 404 returned error can't find the container with id 8079eaecc6f5238a9a35b5f70c66f523d1598acaec34188c2dbb80b4ffcde39c Dec 09 15:48:37 crc kubenswrapper[4894]: I1209 15:48:37.672868 4894 generic.go:334] "Generic (PLEG): container finished" podID="09b8ca6e-7521-4e7f-8272-504d2d7d66d7" containerID="c690018c1d490040c1720247120a6a74843ecabd9c34e41fb997b014f699bf4c" exitCode=0 Dec 09 15:48:37 crc kubenswrapper[4894]: I1209 15:48:37.673287 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-a57d-account-create-update-8xxg4" event={"ID":"09b8ca6e-7521-4e7f-8272-504d2d7d66d7","Type":"ContainerDied","Data":"c690018c1d490040c1720247120a6a74843ecabd9c34e41fb997b014f699bf4c"} Dec 09 15:48:37 crc kubenswrapper[4894]: I1209 15:48:37.677214 4894 generic.go:334] "Generic (PLEG): container finished" podID="81e9c778-7d46-47f9-8d63-15662adc4aa3" containerID="2de20c72f91164109d228f1372091a3e7f4059758bc0f656b112da099e0f37d7" exitCode=0 Dec 09 15:48:37 crc kubenswrapper[4894]: I1209 15:48:37.677278 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5da2-account-create-update-wthfq" event={"ID":"81e9c778-7d46-47f9-8d63-15662adc4aa3","Type":"ContainerDied","Data":"2de20c72f91164109d228f1372091a3e7f4059758bc0f656b112da099e0f37d7"} Dec 09 15:48:37 crc kubenswrapper[4894]: I1209 15:48:37.677298 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5da2-account-create-update-wthfq" event={"ID":"81e9c778-7d46-47f9-8d63-15662adc4aa3","Type":"ContainerStarted","Data":"8079eaecc6f5238a9a35b5f70c66f523d1598acaec34188c2dbb80b4ffcde39c"} Dec 09 15:48:37 crc kubenswrapper[4894]: I1209 15:48:37.678615 4894 generic.go:334] "Generic (PLEG): container finished" podID="b191efc2-e3bf-49e8-83d9-7b2e55264863" containerID="9356d569802254937d2a4dccdd6757c5c7d3bf6a4595e2f8335a3ba17b1703c0" exitCode=0 Dec 09 15:48:37 crc kubenswrapper[4894]: I1209 15:48:37.678713 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-wpqxn" event={"ID":"b191efc2-e3bf-49e8-83d9-7b2e55264863","Type":"ContainerDied","Data":"9356d569802254937d2a4dccdd6757c5c7d3bf6a4595e2f8335a3ba17b1703c0"} Dec 09 15:48:37 crc kubenswrapper[4894]: I1209 15:48:37.680212 4894 generic.go:334] "Generic (PLEG): container finished" podID="a754121b-d60b-4107-ac9c-15cc6db44d93" containerID="723467964f41a9aa68ad1cbed609234ca51ea9e01d04854c81afa5b441925f83" exitCode=0 Dec 09 15:48:37 crc kubenswrapper[4894]: I1209 15:48:37.680290 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-pfn9g" event={"ID":"a754121b-d60b-4107-ac9c-15cc6db44d93","Type":"ContainerDied","Data":"723467964f41a9aa68ad1cbed609234ca51ea9e01d04854c81afa5b441925f83"} Dec 09 15:48:37 crc kubenswrapper[4894]: I1209 15:48:37.681963 4894 generic.go:334] "Generic (PLEG): container finished" podID="48cc59c5-cf8b-455b-8f27-beb7398510b3" containerID="aa04b13b268897f23a37a764d1e04f0bd0ecf5b97c3effe0d1495a6f528e6ea6" exitCode=0 Dec 09 15:48:37 crc kubenswrapper[4894]: I1209 15:48:37.682020 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-v6zx2" event={"ID":"48cc59c5-cf8b-455b-8f27-beb7398510b3","Type":"ContainerDied","Data":"aa04b13b268897f23a37a764d1e04f0bd0ecf5b97c3effe0d1495a6f528e6ea6"} Dec 09 15:48:37 crc kubenswrapper[4894]: I1209 15:48:37.682040 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-v6zx2" event={"ID":"48cc59c5-cf8b-455b-8f27-beb7398510b3","Type":"ContainerStarted","Data":"c71acd0df855104137898aea030c6a8a34375e3c25b41bc9997c7f33519dc302"} Dec 09 15:48:37 crc kubenswrapper[4894]: I1209 15:48:37.684594 4894 generic.go:334] "Generic (PLEG): container finished" podID="946604b5-e779-4579-a593-f46a7c4e3e5e" containerID="d66885ed9842006da476df221f882b744f4f5891d2fe2887ce568f7d6bf2c236" exitCode=0 Dec 09 15:48:37 crc kubenswrapper[4894]: I1209 15:48:37.684709 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-fc79-account-create-update-ts2zw" event={"ID":"946604b5-e779-4579-a593-f46a7c4e3e5e","Type":"ContainerDied","Data":"d66885ed9842006da476df221f882b744f4f5891d2fe2887ce568f7d6bf2c236"} Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.261846 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vk5h6-config-hlk99" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.390540 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa46f14a-27b9-4d99-aeed-63dc95f536ee-scripts\") pod \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.390588 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swwz5\" (UniqueName: \"kubernetes.io/projected/aa46f14a-27b9-4d99-aeed-63dc95f536ee-kube-api-access-swwz5\") pod \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.390646 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/aa46f14a-27b9-4d99-aeed-63dc95f536ee-var-log-ovn\") pod \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.390743 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/aa46f14a-27b9-4d99-aeed-63dc95f536ee-var-run-ovn\") pod \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.390765 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aa46f14a-27b9-4d99-aeed-63dc95f536ee-var-run\") pod \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.390788 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/aa46f14a-27b9-4d99-aeed-63dc95f536ee-additional-scripts\") pod \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\" (UID: \"aa46f14a-27b9-4d99-aeed-63dc95f536ee\") " Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.390841 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa46f14a-27b9-4d99-aeed-63dc95f536ee-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "aa46f14a-27b9-4d99-aeed-63dc95f536ee" (UID: "aa46f14a-27b9-4d99-aeed-63dc95f536ee"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.390855 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa46f14a-27b9-4d99-aeed-63dc95f536ee-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "aa46f14a-27b9-4d99-aeed-63dc95f536ee" (UID: "aa46f14a-27b9-4d99-aeed-63dc95f536ee"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.390889 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/aa46f14a-27b9-4d99-aeed-63dc95f536ee-var-run" (OuterVolumeSpecName: "var-run") pod "aa46f14a-27b9-4d99-aeed-63dc95f536ee" (UID: "aa46f14a-27b9-4d99-aeed-63dc95f536ee"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.391427 4894 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/aa46f14a-27b9-4d99-aeed-63dc95f536ee-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.391456 4894 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/aa46f14a-27b9-4d99-aeed-63dc95f536ee-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.391469 4894 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/aa46f14a-27b9-4d99-aeed-63dc95f536ee-var-run\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.391460 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa46f14a-27b9-4d99-aeed-63dc95f536ee-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "aa46f14a-27b9-4d99-aeed-63dc95f536ee" (UID: "aa46f14a-27b9-4d99-aeed-63dc95f536ee"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.391787 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/aa46f14a-27b9-4d99-aeed-63dc95f536ee-scripts" (OuterVolumeSpecName: "scripts") pod "aa46f14a-27b9-4d99-aeed-63dc95f536ee" (UID: "aa46f14a-27b9-4d99-aeed-63dc95f536ee"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.400864 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/aa46f14a-27b9-4d99-aeed-63dc95f536ee-kube-api-access-swwz5" (OuterVolumeSpecName: "kube-api-access-swwz5") pod "aa46f14a-27b9-4d99-aeed-63dc95f536ee" (UID: "aa46f14a-27b9-4d99-aeed-63dc95f536ee"). InnerVolumeSpecName "kube-api-access-swwz5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.492602 4894 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/aa46f14a-27b9-4d99-aeed-63dc95f536ee-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.492662 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/aa46f14a-27b9-4d99-aeed-63dc95f536ee-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.492678 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swwz5\" (UniqueName: \"kubernetes.io/projected/aa46f14a-27b9-4d99-aeed-63dc95f536ee-kube-api-access-swwz5\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.697078 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vk5h6-config-hlk99" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.697121 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vk5h6-config-hlk99" event={"ID":"aa46f14a-27b9-4d99-aeed-63dc95f536ee","Type":"ContainerDied","Data":"54ec4e5ece5839717f6d174f6298505c1a9eb9a44ef1e416139685121b22d4c2"} Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.697201 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="54ec4e5ece5839717f6d174f6298505c1a9eb9a44ef1e416139685121b22d4c2" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.733489 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-vk5h6-config-hlk99"] Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.744538 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-vk5h6-config-hlk99"] Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.863654 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-vk5h6-config-7b92f"] Dec 09 15:48:38 crc kubenswrapper[4894]: E1209 15:48:38.864040 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="aa46f14a-27b9-4d99-aeed-63dc95f536ee" containerName="ovn-config" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.864054 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="aa46f14a-27b9-4d99-aeed-63dc95f536ee" containerName="ovn-config" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.864260 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="aa46f14a-27b9-4d99-aeed-63dc95f536ee" containerName="ovn-config" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.864816 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vk5h6-config-7b92f" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.867683 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.884207 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vk5h6-config-7b92f"] Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.998521 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c681f297-4c32-4393-8acc-6c3f4fc34ead-var-log-ovn\") pod \"ovn-controller-vk5h6-config-7b92f\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " pod="openstack/ovn-controller-vk5h6-config-7b92f" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.998577 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c681f297-4c32-4393-8acc-6c3f4fc34ead-additional-scripts\") pod \"ovn-controller-vk5h6-config-7b92f\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " pod="openstack/ovn-controller-vk5h6-config-7b92f" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.998608 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j89sj\" (UniqueName: \"kubernetes.io/projected/c681f297-4c32-4393-8acc-6c3f4fc34ead-kube-api-access-j89sj\") pod \"ovn-controller-vk5h6-config-7b92f\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " pod="openstack/ovn-controller-vk5h6-config-7b92f" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.998661 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c681f297-4c32-4393-8acc-6c3f4fc34ead-var-run\") pod \"ovn-controller-vk5h6-config-7b92f\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " pod="openstack/ovn-controller-vk5h6-config-7b92f" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.998727 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c681f297-4c32-4393-8acc-6c3f4fc34ead-var-run-ovn\") pod \"ovn-controller-vk5h6-config-7b92f\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " pod="openstack/ovn-controller-vk5h6-config-7b92f" Dec 09 15:48:38 crc kubenswrapper[4894]: I1209 15:48:38.998844 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c681f297-4c32-4393-8acc-6c3f4fc34ead-scripts\") pod \"ovn-controller-vk5h6-config-7b92f\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " pod="openstack/ovn-controller-vk5h6-config-7b92f" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.099850 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c681f297-4c32-4393-8acc-6c3f4fc34ead-var-run-ovn\") pod \"ovn-controller-vk5h6-config-7b92f\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " pod="openstack/ovn-controller-vk5h6-config-7b92f" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.101630 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c681f297-4c32-4393-8acc-6c3f4fc34ead-scripts\") pod \"ovn-controller-vk5h6-config-7b92f\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " pod="openstack/ovn-controller-vk5h6-config-7b92f" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.101901 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c681f297-4c32-4393-8acc-6c3f4fc34ead-var-log-ovn\") pod \"ovn-controller-vk5h6-config-7b92f\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " pod="openstack/ovn-controller-vk5h6-config-7b92f" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.101934 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c681f297-4c32-4393-8acc-6c3f4fc34ead-additional-scripts\") pod \"ovn-controller-vk5h6-config-7b92f\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " pod="openstack/ovn-controller-vk5h6-config-7b92f" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.101952 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j89sj\" (UniqueName: \"kubernetes.io/projected/c681f297-4c32-4393-8acc-6c3f4fc34ead-kube-api-access-j89sj\") pod \"ovn-controller-vk5h6-config-7b92f\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " pod="openstack/ovn-controller-vk5h6-config-7b92f" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.101983 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c681f297-4c32-4393-8acc-6c3f4fc34ead-var-run\") pod \"ovn-controller-vk5h6-config-7b92f\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " pod="openstack/ovn-controller-vk5h6-config-7b92f" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.102920 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c681f297-4c32-4393-8acc-6c3f4fc34ead-var-run-ovn\") pod \"ovn-controller-vk5h6-config-7b92f\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " pod="openstack/ovn-controller-vk5h6-config-7b92f" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.105512 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c681f297-4c32-4393-8acc-6c3f4fc34ead-scripts\") pod \"ovn-controller-vk5h6-config-7b92f\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " pod="openstack/ovn-controller-vk5h6-config-7b92f" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.107608 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c681f297-4c32-4393-8acc-6c3f4fc34ead-var-run\") pod \"ovn-controller-vk5h6-config-7b92f\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " pod="openstack/ovn-controller-vk5h6-config-7b92f" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.107946 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c681f297-4c32-4393-8acc-6c3f4fc34ead-additional-scripts\") pod \"ovn-controller-vk5h6-config-7b92f\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " pod="openstack/ovn-controller-vk5h6-config-7b92f" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.110480 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c681f297-4c32-4393-8acc-6c3f4fc34ead-var-log-ovn\") pod \"ovn-controller-vk5h6-config-7b92f\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " pod="openstack/ovn-controller-vk5h6-config-7b92f" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.136811 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j89sj\" (UniqueName: \"kubernetes.io/projected/c681f297-4c32-4393-8acc-6c3f4fc34ead-kube-api-access-j89sj\") pod \"ovn-controller-vk5h6-config-7b92f\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " pod="openstack/ovn-controller-vk5h6-config-7b92f" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.149892 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-pfn9g" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.181368 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vk5h6-config-7b92f" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.304119 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a754121b-d60b-4107-ac9c-15cc6db44d93-operator-scripts\") pod \"a754121b-d60b-4107-ac9c-15cc6db44d93\" (UID: \"a754121b-d60b-4107-ac9c-15cc6db44d93\") " Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.304204 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqmhp\" (UniqueName: \"kubernetes.io/projected/a754121b-d60b-4107-ac9c-15cc6db44d93-kube-api-access-fqmhp\") pod \"a754121b-d60b-4107-ac9c-15cc6db44d93\" (UID: \"a754121b-d60b-4107-ac9c-15cc6db44d93\") " Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.305548 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a754121b-d60b-4107-ac9c-15cc6db44d93-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a754121b-d60b-4107-ac9c-15cc6db44d93" (UID: "a754121b-d60b-4107-ac9c-15cc6db44d93"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.308979 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a754121b-d60b-4107-ac9c-15cc6db44d93-kube-api-access-fqmhp" (OuterVolumeSpecName: "kube-api-access-fqmhp") pod "a754121b-d60b-4107-ac9c-15cc6db44d93" (UID: "a754121b-d60b-4107-ac9c-15cc6db44d93"). InnerVolumeSpecName "kube-api-access-fqmhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.365853 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-fc79-account-create-update-ts2zw" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.369560 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-v6zx2" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.396076 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-wpqxn" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.398919 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-a57d-account-create-update-8xxg4" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.403323 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5da2-account-create-update-wthfq" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.424870 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a754121b-d60b-4107-ac9c-15cc6db44d93-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.425234 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqmhp\" (UniqueName: \"kubernetes.io/projected/a754121b-d60b-4107-ac9c-15cc6db44d93-kube-api-access-fqmhp\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.490498 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-vk5h6-config-7b92f"] Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.526121 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggl5x\" (UniqueName: \"kubernetes.io/projected/b191efc2-e3bf-49e8-83d9-7b2e55264863-kube-api-access-ggl5x\") pod \"b191efc2-e3bf-49e8-83d9-7b2e55264863\" (UID: \"b191efc2-e3bf-49e8-83d9-7b2e55264863\") " Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.526219 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w2vb4\" (UniqueName: \"kubernetes.io/projected/946604b5-e779-4579-a593-f46a7c4e3e5e-kube-api-access-w2vb4\") pod \"946604b5-e779-4579-a593-f46a7c4e3e5e\" (UID: \"946604b5-e779-4579-a593-f46a7c4e3e5e\") " Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.526244 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dl7xf\" (UniqueName: \"kubernetes.io/projected/48cc59c5-cf8b-455b-8f27-beb7398510b3-kube-api-access-dl7xf\") pod \"48cc59c5-cf8b-455b-8f27-beb7398510b3\" (UID: \"48cc59c5-cf8b-455b-8f27-beb7398510b3\") " Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.526298 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/946604b5-e779-4579-a593-f46a7c4e3e5e-operator-scripts\") pod \"946604b5-e779-4579-a593-f46a7c4e3e5e\" (UID: \"946604b5-e779-4579-a593-f46a7c4e3e5e\") " Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.526332 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrhc6\" (UniqueName: \"kubernetes.io/projected/09b8ca6e-7521-4e7f-8272-504d2d7d66d7-kube-api-access-xrhc6\") pod \"09b8ca6e-7521-4e7f-8272-504d2d7d66d7\" (UID: \"09b8ca6e-7521-4e7f-8272-504d2d7d66d7\") " Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.526359 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09b8ca6e-7521-4e7f-8272-504d2d7d66d7-operator-scripts\") pod \"09b8ca6e-7521-4e7f-8272-504d2d7d66d7\" (UID: \"09b8ca6e-7521-4e7f-8272-504d2d7d66d7\") " Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.526391 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48cc59c5-cf8b-455b-8f27-beb7398510b3-operator-scripts\") pod \"48cc59c5-cf8b-455b-8f27-beb7398510b3\" (UID: \"48cc59c5-cf8b-455b-8f27-beb7398510b3\") " Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.526408 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gp2d9\" (UniqueName: \"kubernetes.io/projected/81e9c778-7d46-47f9-8d63-15662adc4aa3-kube-api-access-gp2d9\") pod \"81e9c778-7d46-47f9-8d63-15662adc4aa3\" (UID: \"81e9c778-7d46-47f9-8d63-15662adc4aa3\") " Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.526424 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b191efc2-e3bf-49e8-83d9-7b2e55264863-operator-scripts\") pod \"b191efc2-e3bf-49e8-83d9-7b2e55264863\" (UID: \"b191efc2-e3bf-49e8-83d9-7b2e55264863\") " Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.526447 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81e9c778-7d46-47f9-8d63-15662adc4aa3-operator-scripts\") pod \"81e9c778-7d46-47f9-8d63-15662adc4aa3\" (UID: \"81e9c778-7d46-47f9-8d63-15662adc4aa3\") " Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.526860 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09b8ca6e-7521-4e7f-8272-504d2d7d66d7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "09b8ca6e-7521-4e7f-8272-504d2d7d66d7" (UID: "09b8ca6e-7521-4e7f-8272-504d2d7d66d7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.526927 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/946604b5-e779-4579-a593-f46a7c4e3e5e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "946604b5-e779-4579-a593-f46a7c4e3e5e" (UID: "946604b5-e779-4579-a593-f46a7c4e3e5e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.527138 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/81e9c778-7d46-47f9-8d63-15662adc4aa3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "81e9c778-7d46-47f9-8d63-15662adc4aa3" (UID: "81e9c778-7d46-47f9-8d63-15662adc4aa3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.527562 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b191efc2-e3bf-49e8-83d9-7b2e55264863-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b191efc2-e3bf-49e8-83d9-7b2e55264863" (UID: "b191efc2-e3bf-49e8-83d9-7b2e55264863"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.527558 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/48cc59c5-cf8b-455b-8f27-beb7398510b3-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "48cc59c5-cf8b-455b-8f27-beb7398510b3" (UID: "48cc59c5-cf8b-455b-8f27-beb7398510b3"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.530482 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/81e9c778-7d46-47f9-8d63-15662adc4aa3-kube-api-access-gp2d9" (OuterVolumeSpecName: "kube-api-access-gp2d9") pod "81e9c778-7d46-47f9-8d63-15662adc4aa3" (UID: "81e9c778-7d46-47f9-8d63-15662adc4aa3"). InnerVolumeSpecName "kube-api-access-gp2d9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.531390 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/946604b5-e779-4579-a593-f46a7c4e3e5e-kube-api-access-w2vb4" (OuterVolumeSpecName: "kube-api-access-w2vb4") pod "946604b5-e779-4579-a593-f46a7c4e3e5e" (UID: "946604b5-e779-4579-a593-f46a7c4e3e5e"). InnerVolumeSpecName "kube-api-access-w2vb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.532015 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48cc59c5-cf8b-455b-8f27-beb7398510b3-kube-api-access-dl7xf" (OuterVolumeSpecName: "kube-api-access-dl7xf") pod "48cc59c5-cf8b-455b-8f27-beb7398510b3" (UID: "48cc59c5-cf8b-455b-8f27-beb7398510b3"). InnerVolumeSpecName "kube-api-access-dl7xf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.532369 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09b8ca6e-7521-4e7f-8272-504d2d7d66d7-kube-api-access-xrhc6" (OuterVolumeSpecName: "kube-api-access-xrhc6") pod "09b8ca6e-7521-4e7f-8272-504d2d7d66d7" (UID: "09b8ca6e-7521-4e7f-8272-504d2d7d66d7"). InnerVolumeSpecName "kube-api-access-xrhc6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.532563 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b191efc2-e3bf-49e8-83d9-7b2e55264863-kube-api-access-ggl5x" (OuterVolumeSpecName: "kube-api-access-ggl5x") pod "b191efc2-e3bf-49e8-83d9-7b2e55264863" (UID: "b191efc2-e3bf-49e8-83d9-7b2e55264863"). InnerVolumeSpecName "kube-api-access-ggl5x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.628756 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/09b8ca6e-7521-4e7f-8272-504d2d7d66d7-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.628795 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/48cc59c5-cf8b-455b-8f27-beb7398510b3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.628808 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gp2d9\" (UniqueName: \"kubernetes.io/projected/81e9c778-7d46-47f9-8d63-15662adc4aa3-kube-api-access-gp2d9\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.628821 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b191efc2-e3bf-49e8-83d9-7b2e55264863-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.628832 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/81e9c778-7d46-47f9-8d63-15662adc4aa3-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.628844 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggl5x\" (UniqueName: \"kubernetes.io/projected/b191efc2-e3bf-49e8-83d9-7b2e55264863-kube-api-access-ggl5x\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.628855 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w2vb4\" (UniqueName: \"kubernetes.io/projected/946604b5-e779-4579-a593-f46a7c4e3e5e-kube-api-access-w2vb4\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.628865 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dl7xf\" (UniqueName: \"kubernetes.io/projected/48cc59c5-cf8b-455b-8f27-beb7398510b3-kube-api-access-dl7xf\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.628888 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/946604b5-e779-4579-a593-f46a7c4e3e5e-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.628899 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrhc6\" (UniqueName: \"kubernetes.io/projected/09b8ca6e-7521-4e7f-8272-504d2d7d66d7-kube-api-access-xrhc6\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.707303 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5da2-account-create-update-wthfq" event={"ID":"81e9c778-7d46-47f9-8d63-15662adc4aa3","Type":"ContainerDied","Data":"8079eaecc6f5238a9a35b5f70c66f523d1598acaec34188c2dbb80b4ffcde39c"} Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.707343 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8079eaecc6f5238a9a35b5f70c66f523d1598acaec34188c2dbb80b4ffcde39c" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.707346 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5da2-account-create-update-wthfq" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.708918 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-wpqxn" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.708936 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-wpqxn" event={"ID":"b191efc2-e3bf-49e8-83d9-7b2e55264863","Type":"ContainerDied","Data":"1a961f4663ff5fd098c4610a9ffff4d556636698ccbc761c2ffde5eabdbe33db"} Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.708981 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1a961f4663ff5fd098c4610a9ffff4d556636698ccbc761c2ffde5eabdbe33db" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.715094 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-pfn9g" event={"ID":"a754121b-d60b-4107-ac9c-15cc6db44d93","Type":"ContainerDied","Data":"7d41feca8c72d8a58b64feec1609138aa5df13c8659d94cf6fafd7fc61f9c7d5"} Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.715111 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d41feca8c72d8a58b64feec1609138aa5df13c8659d94cf6fafd7fc61f9c7d5" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.715174 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-pfn9g" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.722832 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-v6zx2" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.722839 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-v6zx2" event={"ID":"48cc59c5-cf8b-455b-8f27-beb7398510b3","Type":"ContainerDied","Data":"c71acd0df855104137898aea030c6a8a34375e3c25b41bc9997c7f33519dc302"} Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.723215 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c71acd0df855104137898aea030c6a8a34375e3c25b41bc9997c7f33519dc302" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.724704 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-fc79-account-create-update-ts2zw" event={"ID":"946604b5-e779-4579-a593-f46a7c4e3e5e","Type":"ContainerDied","Data":"70ba905c2a3749be73032e64b0960018e62eba975c552ab97f6496c9bdaad786"} Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.724752 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="70ba905c2a3749be73032e64b0960018e62eba975c552ab97f6496c9bdaad786" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.724839 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-fc79-account-create-update-ts2zw" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.726796 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-a57d-account-create-update-8xxg4" event={"ID":"09b8ca6e-7521-4e7f-8272-504d2d7d66d7","Type":"ContainerDied","Data":"083214a4b8bca511aade9bc4f97cc155e08a9e4e7230cd3072c92e617733d305"} Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.726824 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="083214a4b8bca511aade9bc4f97cc155e08a9e4e7230cd3072c92e617733d305" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.726886 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-a57d-account-create-update-8xxg4" Dec 09 15:48:39 crc kubenswrapper[4894]: I1209 15:48:39.728165 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vk5h6-config-7b92f" event={"ID":"c681f297-4c32-4393-8acc-6c3f4fc34ead","Type":"ContainerStarted","Data":"c33c24733804411edaf7077b8fd9ae6b6f6691257b4f87022ac84c6ed173363d"} Dec 09 15:48:40 crc kubenswrapper[4894]: I1209 15:48:40.123044 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="aa46f14a-27b9-4d99-aeed-63dc95f536ee" path="/var/lib/kubelet/pods/aa46f14a-27b9-4d99-aeed-63dc95f536ee/volumes" Dec 09 15:48:40 crc kubenswrapper[4894]: I1209 15:48:40.737137 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vk5h6-config-7b92f" event={"ID":"c681f297-4c32-4393-8acc-6c3f4fc34ead","Type":"ContainerStarted","Data":"d64cbbd062255979137593fdda62ec306b7b9a73e617f4bf24a5728dd1d475c3"} Dec 09 15:48:41 crc kubenswrapper[4894]: I1209 15:48:41.747901 4894 generic.go:334] "Generic (PLEG): container finished" podID="c681f297-4c32-4393-8acc-6c3f4fc34ead" containerID="d64cbbd062255979137593fdda62ec306b7b9a73e617f4bf24a5728dd1d475c3" exitCode=0 Dec 09 15:48:41 crc kubenswrapper[4894]: I1209 15:48:41.747946 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vk5h6-config-7b92f" event={"ID":"c681f297-4c32-4393-8acc-6c3f4fc34ead","Type":"ContainerDied","Data":"d64cbbd062255979137593fdda62ec306b7b9a73e617f4bf24a5728dd1d475c3"} Dec 09 15:48:42 crc kubenswrapper[4894]: I1209 15:48:42.275789 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" Dec 09 15:48:42 crc kubenswrapper[4894]: I1209 15:48:42.328917 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-586b989cdc-946tp"] Dec 09 15:48:42 crc kubenswrapper[4894]: I1209 15:48:42.329562 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-586b989cdc-946tp" podUID="640684c9-0085-424c-a19d-3510625cef05" containerName="dnsmasq-dns" containerID="cri-o://073c895aeb5d86d84883745aa5ee2682a9db1f2d05a9348e1633116b21b984d0" gracePeriod=10 Dec 09 15:48:45 crc kubenswrapper[4894]: I1209 15:48:45.747573 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vk5h6-config-7b92f" Dec 09 15:48:45 crc kubenswrapper[4894]: I1209 15:48:45.782417 4894 generic.go:334] "Generic (PLEG): container finished" podID="640684c9-0085-424c-a19d-3510625cef05" containerID="073c895aeb5d86d84883745aa5ee2682a9db1f2d05a9348e1633116b21b984d0" exitCode=0 Dec 09 15:48:45 crc kubenswrapper[4894]: I1209 15:48:45.782512 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586b989cdc-946tp" event={"ID":"640684c9-0085-424c-a19d-3510625cef05","Type":"ContainerDied","Data":"073c895aeb5d86d84883745aa5ee2682a9db1f2d05a9348e1633116b21b984d0"} Dec 09 15:48:45 crc kubenswrapper[4894]: I1209 15:48:45.784273 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-vk5h6-config-7b92f" event={"ID":"c681f297-4c32-4393-8acc-6c3f4fc34ead","Type":"ContainerDied","Data":"c33c24733804411edaf7077b8fd9ae6b6f6691257b4f87022ac84c6ed173363d"} Dec 09 15:48:45 crc kubenswrapper[4894]: I1209 15:48:45.784309 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c33c24733804411edaf7077b8fd9ae6b6f6691257b4f87022ac84c6ed173363d" Dec 09 15:48:45 crc kubenswrapper[4894]: I1209 15:48:45.784316 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-vk5h6-config-7b92f" Dec 09 15:48:45 crc kubenswrapper[4894]: I1209 15:48:45.899142 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586b989cdc-946tp" Dec 09 15:48:45 crc kubenswrapper[4894]: I1209 15:48:45.945596 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c681f297-4c32-4393-8acc-6c3f4fc34ead-scripts\") pod \"c681f297-4c32-4393-8acc-6c3f4fc34ead\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " Dec 09 15:48:45 crc kubenswrapper[4894]: I1209 15:48:45.945677 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c681f297-4c32-4393-8acc-6c3f4fc34ead-additional-scripts\") pod \"c681f297-4c32-4393-8acc-6c3f4fc34ead\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " Dec 09 15:48:45 crc kubenswrapper[4894]: I1209 15:48:45.945734 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c681f297-4c32-4393-8acc-6c3f4fc34ead-var-run\") pod \"c681f297-4c32-4393-8acc-6c3f4fc34ead\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " Dec 09 15:48:45 crc kubenswrapper[4894]: I1209 15:48:45.945783 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j89sj\" (UniqueName: \"kubernetes.io/projected/c681f297-4c32-4393-8acc-6c3f4fc34ead-kube-api-access-j89sj\") pod \"c681f297-4c32-4393-8acc-6c3f4fc34ead\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " Dec 09 15:48:45 crc kubenswrapper[4894]: I1209 15:48:45.945799 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c681f297-4c32-4393-8acc-6c3f4fc34ead-var-run-ovn\") pod \"c681f297-4c32-4393-8acc-6c3f4fc34ead\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " Dec 09 15:48:45 crc kubenswrapper[4894]: I1209 15:48:45.945894 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c681f297-4c32-4393-8acc-6c3f4fc34ead-var-log-ovn\") pod \"c681f297-4c32-4393-8acc-6c3f4fc34ead\" (UID: \"c681f297-4c32-4393-8acc-6c3f4fc34ead\") " Dec 09 15:48:45 crc kubenswrapper[4894]: I1209 15:48:45.946239 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c681f297-4c32-4393-8acc-6c3f4fc34ead-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "c681f297-4c32-4393-8acc-6c3f4fc34ead" (UID: "c681f297-4c32-4393-8acc-6c3f4fc34ead"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:48:45 crc kubenswrapper[4894]: I1209 15:48:45.946740 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c681f297-4c32-4393-8acc-6c3f4fc34ead-var-run" (OuterVolumeSpecName: "var-run") pod "c681f297-4c32-4393-8acc-6c3f4fc34ead" (UID: "c681f297-4c32-4393-8acc-6c3f4fc34ead"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:48:45 crc kubenswrapper[4894]: I1209 15:48:45.947047 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c681f297-4c32-4393-8acc-6c3f4fc34ead-scripts" (OuterVolumeSpecName: "scripts") pod "c681f297-4c32-4393-8acc-6c3f4fc34ead" (UID: "c681f297-4c32-4393-8acc-6c3f4fc34ead"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:45 crc kubenswrapper[4894]: I1209 15:48:45.947099 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/c681f297-4c32-4393-8acc-6c3f4fc34ead-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "c681f297-4c32-4393-8acc-6c3f4fc34ead" (UID: "c681f297-4c32-4393-8acc-6c3f4fc34ead"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:48:45 crc kubenswrapper[4894]: I1209 15:48:45.947671 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c681f297-4c32-4393-8acc-6c3f4fc34ead-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "c681f297-4c32-4393-8acc-6c3f4fc34ead" (UID: "c681f297-4c32-4393-8acc-6c3f4fc34ead"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:45 crc kubenswrapper[4894]: I1209 15:48:45.951173 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c681f297-4c32-4393-8acc-6c3f4fc34ead-kube-api-access-j89sj" (OuterVolumeSpecName: "kube-api-access-j89sj") pod "c681f297-4c32-4393-8acc-6c3f4fc34ead" (UID: "c681f297-4c32-4393-8acc-6c3f4fc34ead"). InnerVolumeSpecName "kube-api-access-j89sj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.047629 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nncbx\" (UniqueName: \"kubernetes.io/projected/640684c9-0085-424c-a19d-3510625cef05-kube-api-access-nncbx\") pod \"640684c9-0085-424c-a19d-3510625cef05\" (UID: \"640684c9-0085-424c-a19d-3510625cef05\") " Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.047709 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-ovsdbserver-sb\") pod \"640684c9-0085-424c-a19d-3510625cef05\" (UID: \"640684c9-0085-424c-a19d-3510625cef05\") " Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.050629 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/640684c9-0085-424c-a19d-3510625cef05-kube-api-access-nncbx" (OuterVolumeSpecName: "kube-api-access-nncbx") pod "640684c9-0085-424c-a19d-3510625cef05" (UID: "640684c9-0085-424c-a19d-3510625cef05"). InnerVolumeSpecName "kube-api-access-nncbx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.051764 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-ovsdbserver-nb\") pod \"640684c9-0085-424c-a19d-3510625cef05\" (UID: \"640684c9-0085-424c-a19d-3510625cef05\") " Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.051840 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-dns-svc\") pod \"640684c9-0085-424c-a19d-3510625cef05\" (UID: \"640684c9-0085-424c-a19d-3510625cef05\") " Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.051858 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-config\") pod \"640684c9-0085-424c-a19d-3510625cef05\" (UID: \"640684c9-0085-424c-a19d-3510625cef05\") " Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.052380 4894 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/c681f297-4c32-4393-8acc-6c3f4fc34ead-var-run\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.052396 4894 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/c681f297-4c32-4393-8acc-6c3f4fc34ead-var-run-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.052409 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j89sj\" (UniqueName: \"kubernetes.io/projected/c681f297-4c32-4393-8acc-6c3f4fc34ead-kube-api-access-j89sj\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.052422 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nncbx\" (UniqueName: \"kubernetes.io/projected/640684c9-0085-424c-a19d-3510625cef05-kube-api-access-nncbx\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.052433 4894 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/c681f297-4c32-4393-8acc-6c3f4fc34ead-var-log-ovn\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.052443 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/c681f297-4c32-4393-8acc-6c3f4fc34ead-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.052453 4894 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/c681f297-4c32-4393-8acc-6c3f4fc34ead-additional-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.085230 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "640684c9-0085-424c-a19d-3510625cef05" (UID: "640684c9-0085-424c-a19d-3510625cef05"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.087180 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "640684c9-0085-424c-a19d-3510625cef05" (UID: "640684c9-0085-424c-a19d-3510625cef05"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.088319 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "640684c9-0085-424c-a19d-3510625cef05" (UID: "640684c9-0085-424c-a19d-3510625cef05"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.093707 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-config" (OuterVolumeSpecName: "config") pod "640684c9-0085-424c-a19d-3510625cef05" (UID: "640684c9-0085-424c-a19d-3510625cef05"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.161284 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.161753 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.161792 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.161816 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/640684c9-0085-424c-a19d-3510625cef05-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.805144 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-npqvf" event={"ID":"87072774-8453-49d2-a355-3388179d06dc","Type":"ContainerStarted","Data":"0abfc9e8d74dfb18b88220cbd3c45dca30fc929eacadd325f0ba4baa46b1b3e9"} Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.808581 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-586b989cdc-946tp" event={"ID":"640684c9-0085-424c-a19d-3510625cef05","Type":"ContainerDied","Data":"6213b765214edb517bb6d70e8fbc638808980173618f278fa8f9837c25410b75"} Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.808658 4894 scope.go:117] "RemoveContainer" containerID="073c895aeb5d86d84883745aa5ee2682a9db1f2d05a9348e1633116b21b984d0" Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.808852 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-586b989cdc-946tp" Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.818810 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-vk5h6-config-7b92f"] Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.829687 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-vk5h6-config-7b92f"] Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.845818 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-npqvf" podStartSLOduration=2.673943146 podStartE2EDuration="11.84579733s" podCreationTimestamp="2025-12-09 15:48:35 +0000 UTC" firstStartedPulling="2025-12-09 15:48:36.558076461 +0000 UTC m=+1010.877287130" lastFinishedPulling="2025-12-09 15:48:45.729930645 +0000 UTC m=+1020.049141314" observedRunningTime="2025-12-09 15:48:46.824745599 +0000 UTC m=+1021.143956278" watchObservedRunningTime="2025-12-09 15:48:46.84579733 +0000 UTC m=+1021.165007999" Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.847597 4894 scope.go:117] "RemoveContainer" containerID="c17f44953076f2c6ef3d3fa75facd18c28928a9d052516faa5cb846118da9409" Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.856872 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-586b989cdc-946tp"] Dec 09 15:48:46 crc kubenswrapper[4894]: I1209 15:48:46.863051 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-586b989cdc-946tp"] Dec 09 15:48:48 crc kubenswrapper[4894]: I1209 15:48:48.117014 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="640684c9-0085-424c-a19d-3510625cef05" path="/var/lib/kubelet/pods/640684c9-0085-424c-a19d-3510625cef05/volumes" Dec 09 15:48:48 crc kubenswrapper[4894]: I1209 15:48:48.118869 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c681f297-4c32-4393-8acc-6c3f4fc34ead" path="/var/lib/kubelet/pods/c681f297-4c32-4393-8acc-6c3f4fc34ead/volumes" Dec 09 15:48:56 crc kubenswrapper[4894]: I1209 15:48:56.889354 4894 generic.go:334] "Generic (PLEG): container finished" podID="87072774-8453-49d2-a355-3388179d06dc" containerID="0abfc9e8d74dfb18b88220cbd3c45dca30fc929eacadd325f0ba4baa46b1b3e9" exitCode=0 Dec 09 15:48:56 crc kubenswrapper[4894]: I1209 15:48:56.889461 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-npqvf" event={"ID":"87072774-8453-49d2-a355-3388179d06dc","Type":"ContainerDied","Data":"0abfc9e8d74dfb18b88220cbd3c45dca30fc929eacadd325f0ba4baa46b1b3e9"} Dec 09 15:48:58 crc kubenswrapper[4894]: I1209 15:48:58.226688 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-npqvf" Dec 09 15:48:58 crc kubenswrapper[4894]: I1209 15:48:58.334389 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8xxzk\" (UniqueName: \"kubernetes.io/projected/87072774-8453-49d2-a355-3388179d06dc-kube-api-access-8xxzk\") pod \"87072774-8453-49d2-a355-3388179d06dc\" (UID: \"87072774-8453-49d2-a355-3388179d06dc\") " Dec 09 15:48:58 crc kubenswrapper[4894]: I1209 15:48:58.334571 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87072774-8453-49d2-a355-3388179d06dc-combined-ca-bundle\") pod \"87072774-8453-49d2-a355-3388179d06dc\" (UID: \"87072774-8453-49d2-a355-3388179d06dc\") " Dec 09 15:48:58 crc kubenswrapper[4894]: I1209 15:48:58.334729 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87072774-8453-49d2-a355-3388179d06dc-config-data\") pod \"87072774-8453-49d2-a355-3388179d06dc\" (UID: \"87072774-8453-49d2-a355-3388179d06dc\") " Dec 09 15:48:58 crc kubenswrapper[4894]: I1209 15:48:58.345929 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87072774-8453-49d2-a355-3388179d06dc-kube-api-access-8xxzk" (OuterVolumeSpecName: "kube-api-access-8xxzk") pod "87072774-8453-49d2-a355-3388179d06dc" (UID: "87072774-8453-49d2-a355-3388179d06dc"). InnerVolumeSpecName "kube-api-access-8xxzk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:48:58 crc kubenswrapper[4894]: I1209 15:48:58.360495 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87072774-8453-49d2-a355-3388179d06dc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "87072774-8453-49d2-a355-3388179d06dc" (UID: "87072774-8453-49d2-a355-3388179d06dc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:48:58 crc kubenswrapper[4894]: I1209 15:48:58.383519 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87072774-8453-49d2-a355-3388179d06dc-config-data" (OuterVolumeSpecName: "config-data") pod "87072774-8453-49d2-a355-3388179d06dc" (UID: "87072774-8453-49d2-a355-3388179d06dc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:48:58 crc kubenswrapper[4894]: I1209 15:48:58.436337 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87072774-8453-49d2-a355-3388179d06dc-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:58 crc kubenswrapper[4894]: I1209 15:48:58.436369 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8xxzk\" (UniqueName: \"kubernetes.io/projected/87072774-8453-49d2-a355-3388179d06dc-kube-api-access-8xxzk\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:58 crc kubenswrapper[4894]: I1209 15:48:58.436379 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87072774-8453-49d2-a355-3388179d06dc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:48:58 crc kubenswrapper[4894]: I1209 15:48:58.913649 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-npqvf" event={"ID":"87072774-8453-49d2-a355-3388179d06dc","Type":"ContainerDied","Data":"3437db2c3e7023d1f04456ab6d82e8c3ba39ff11ccf2cd533ff549d5cdc4a175"} Dec 09 15:48:58 crc kubenswrapper[4894]: I1209 15:48:58.913906 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3437db2c3e7023d1f04456ab6d82e8c3ba39ff11ccf2cd533ff549d5cdc4a175" Dec 09 15:48:58 crc kubenswrapper[4894]: I1209 15:48:58.913984 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-npqvf" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.169034 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-66dd8c6975-c6tff"] Dec 09 15:48:59 crc kubenswrapper[4894]: E1209 15:48:59.169447 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="640684c9-0085-424c-a19d-3510625cef05" containerName="dnsmasq-dns" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.169465 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="640684c9-0085-424c-a19d-3510625cef05" containerName="dnsmasq-dns" Dec 09 15:48:59 crc kubenswrapper[4894]: E1209 15:48:59.169472 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a754121b-d60b-4107-ac9c-15cc6db44d93" containerName="mariadb-database-create" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.169478 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a754121b-d60b-4107-ac9c-15cc6db44d93" containerName="mariadb-database-create" Dec 09 15:48:59 crc kubenswrapper[4894]: E1209 15:48:59.169515 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48cc59c5-cf8b-455b-8f27-beb7398510b3" containerName="mariadb-database-create" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.169522 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="48cc59c5-cf8b-455b-8f27-beb7398510b3" containerName="mariadb-database-create" Dec 09 15:48:59 crc kubenswrapper[4894]: E1209 15:48:59.169531 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="09b8ca6e-7521-4e7f-8272-504d2d7d66d7" containerName="mariadb-account-create-update" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.169537 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="09b8ca6e-7521-4e7f-8272-504d2d7d66d7" containerName="mariadb-account-create-update" Dec 09 15:48:59 crc kubenswrapper[4894]: E1209 15:48:59.169548 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="81e9c778-7d46-47f9-8d63-15662adc4aa3" containerName="mariadb-account-create-update" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.169554 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="81e9c778-7d46-47f9-8d63-15662adc4aa3" containerName="mariadb-account-create-update" Dec 09 15:48:59 crc kubenswrapper[4894]: E1209 15:48:59.169566 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87072774-8453-49d2-a355-3388179d06dc" containerName="keystone-db-sync" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.169572 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="87072774-8453-49d2-a355-3388179d06dc" containerName="keystone-db-sync" Dec 09 15:48:59 crc kubenswrapper[4894]: E1209 15:48:59.169595 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c681f297-4c32-4393-8acc-6c3f4fc34ead" containerName="ovn-config" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.169603 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c681f297-4c32-4393-8acc-6c3f4fc34ead" containerName="ovn-config" Dec 09 15:48:59 crc kubenswrapper[4894]: E1209 15:48:59.169612 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="640684c9-0085-424c-a19d-3510625cef05" containerName="init" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.169617 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="640684c9-0085-424c-a19d-3510625cef05" containerName="init" Dec 09 15:48:59 crc kubenswrapper[4894]: E1209 15:48:59.169626 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b191efc2-e3bf-49e8-83d9-7b2e55264863" containerName="mariadb-database-create" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.169631 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b191efc2-e3bf-49e8-83d9-7b2e55264863" containerName="mariadb-database-create" Dec 09 15:48:59 crc kubenswrapper[4894]: E1209 15:48:59.169703 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="946604b5-e779-4579-a593-f46a7c4e3e5e" containerName="mariadb-account-create-update" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.169709 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="946604b5-e779-4579-a593-f46a7c4e3e5e" containerName="mariadb-account-create-update" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.169889 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="a754121b-d60b-4107-ac9c-15cc6db44d93" containerName="mariadb-database-create" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.169924 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="946604b5-e779-4579-a593-f46a7c4e3e5e" containerName="mariadb-account-create-update" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.169936 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="640684c9-0085-424c-a19d-3510625cef05" containerName="dnsmasq-dns" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.169943 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="c681f297-4c32-4393-8acc-6c3f4fc34ead" containerName="ovn-config" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.169954 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="48cc59c5-cf8b-455b-8f27-beb7398510b3" containerName="mariadb-database-create" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.169963 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b191efc2-e3bf-49e8-83d9-7b2e55264863" containerName="mariadb-database-create" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.169970 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="09b8ca6e-7521-4e7f-8272-504d2d7d66d7" containerName="mariadb-account-create-update" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.170000 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="81e9c778-7d46-47f9-8d63-15662adc4aa3" containerName="mariadb-account-create-update" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.170010 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="87072774-8453-49d2-a355-3388179d06dc" containerName="keystone-db-sync" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.171707 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66dd8c6975-c6tff" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.187085 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-66dd8c6975-c6tff"] Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.212001 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-274hf"] Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.219958 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-274hf" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.230498 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.230581 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.230865 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.230966 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-g8fmh" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.231027 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.231911 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-274hf"] Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.257283 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-ovsdbserver-nb\") pod \"dnsmasq-dns-66dd8c6975-c6tff\" (UID: \"453feb9e-4bbd-49b8-9719-07f97af54b46\") " pod="openstack/dnsmasq-dns-66dd8c6975-c6tff" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.257327 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-dns-svc\") pod \"dnsmasq-dns-66dd8c6975-c6tff\" (UID: \"453feb9e-4bbd-49b8-9719-07f97af54b46\") " pod="openstack/dnsmasq-dns-66dd8c6975-c6tff" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.257439 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-config\") pod \"dnsmasq-dns-66dd8c6975-c6tff\" (UID: \"453feb9e-4bbd-49b8-9719-07f97af54b46\") " pod="openstack/dnsmasq-dns-66dd8c6975-c6tff" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.257461 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-ovsdbserver-sb\") pod \"dnsmasq-dns-66dd8c6975-c6tff\" (UID: \"453feb9e-4bbd-49b8-9719-07f97af54b46\") " pod="openstack/dnsmasq-dns-66dd8c6975-c6tff" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.257528 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rcx45\" (UniqueName: \"kubernetes.io/projected/453feb9e-4bbd-49b8-9719-07f97af54b46-kube-api-access-rcx45\") pod \"dnsmasq-dns-66dd8c6975-c6tff\" (UID: \"453feb9e-4bbd-49b8-9719-07f97af54b46\") " pod="openstack/dnsmasq-dns-66dd8c6975-c6tff" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.343288 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-55fff446b9-mbvp9"] Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.344911 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55fff446b9-mbvp9" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.358919 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-scripts\") pod \"keystone-bootstrap-274hf\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " pod="openstack/keystone-bootstrap-274hf" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.358964 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-combined-ca-bundle\") pod \"keystone-bootstrap-274hf\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " pod="openstack/keystone-bootstrap-274hf" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.358995 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-credential-keys\") pod \"keystone-bootstrap-274hf\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " pod="openstack/keystone-bootstrap-274hf" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.359023 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rcx45\" (UniqueName: \"kubernetes.io/projected/453feb9e-4bbd-49b8-9719-07f97af54b46-kube-api-access-rcx45\") pod \"dnsmasq-dns-66dd8c6975-c6tff\" (UID: \"453feb9e-4bbd-49b8-9719-07f97af54b46\") " pod="openstack/dnsmasq-dns-66dd8c6975-c6tff" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.359045 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-fernet-keys\") pod \"keystone-bootstrap-274hf\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " pod="openstack/keystone-bootstrap-274hf" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.359073 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-ovsdbserver-nb\") pod \"dnsmasq-dns-66dd8c6975-c6tff\" (UID: \"453feb9e-4bbd-49b8-9719-07f97af54b46\") " pod="openstack/dnsmasq-dns-66dd8c6975-c6tff" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.359091 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-dns-svc\") pod \"dnsmasq-dns-66dd8c6975-c6tff\" (UID: \"453feb9e-4bbd-49b8-9719-07f97af54b46\") " pod="openstack/dnsmasq-dns-66dd8c6975-c6tff" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.359164 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-config\") pod \"dnsmasq-dns-66dd8c6975-c6tff\" (UID: \"453feb9e-4bbd-49b8-9719-07f97af54b46\") " pod="openstack/dnsmasq-dns-66dd8c6975-c6tff" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.359193 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-ovsdbserver-sb\") pod \"dnsmasq-dns-66dd8c6975-c6tff\" (UID: \"453feb9e-4bbd-49b8-9719-07f97af54b46\") " pod="openstack/dnsmasq-dns-66dd8c6975-c6tff" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.359231 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5ldj\" (UniqueName: \"kubernetes.io/projected/ae433ea4-7305-4d95-96b5-da695d0ed504-kube-api-access-t5ldj\") pod \"keystone-bootstrap-274hf\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " pod="openstack/keystone-bootstrap-274hf" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.359254 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-config-data\") pod \"keystone-bootstrap-274hf\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " pod="openstack/keystone-bootstrap-274hf" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.360376 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-ovsdbserver-nb\") pod \"dnsmasq-dns-66dd8c6975-c6tff\" (UID: \"453feb9e-4bbd-49b8-9719-07f97af54b46\") " pod="openstack/dnsmasq-dns-66dd8c6975-c6tff" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.360942 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-dns-svc\") pod \"dnsmasq-dns-66dd8c6975-c6tff\" (UID: \"453feb9e-4bbd-49b8-9719-07f97af54b46\") " pod="openstack/dnsmasq-dns-66dd8c6975-c6tff" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.361475 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-config\") pod \"dnsmasq-dns-66dd8c6975-c6tff\" (UID: \"453feb9e-4bbd-49b8-9719-07f97af54b46\") " pod="openstack/dnsmasq-dns-66dd8c6975-c6tff" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.361726 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-ovsdbserver-sb\") pod \"dnsmasq-dns-66dd8c6975-c6tff\" (UID: \"453feb9e-4bbd-49b8-9719-07f97af54b46\") " pod="openstack/dnsmasq-dns-66dd8c6975-c6tff" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.366800 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon-horizon-dockercfg-bfvkj" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.367428 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"horizon" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.367549 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-config-data" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.367609 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizon-scripts" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.374075 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-55fff446b9-mbvp9"] Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.409309 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rcx45\" (UniqueName: \"kubernetes.io/projected/453feb9e-4bbd-49b8-9719-07f97af54b46-kube-api-access-rcx45\") pod \"dnsmasq-dns-66dd8c6975-c6tff\" (UID: \"453feb9e-4bbd-49b8-9719-07f97af54b46\") " pod="openstack/dnsmasq-dns-66dd8c6975-c6tff" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.460501 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-scripts\") pod \"keystone-bootstrap-274hf\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " pod="openstack/keystone-bootstrap-274hf" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.460550 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-combined-ca-bundle\") pod \"keystone-bootstrap-274hf\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " pod="openstack/keystone-bootstrap-274hf" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.460575 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-credential-keys\") pod \"keystone-bootstrap-274hf\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " pod="openstack/keystone-bootstrap-274hf" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.460600 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff441522-39b1-47e3-be7d-d590baec8889-config-data\") pod \"horizon-55fff446b9-mbvp9\" (UID: \"ff441522-39b1-47e3-be7d-d590baec8889\") " pod="openstack/horizon-55fff446b9-mbvp9" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.460622 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-fernet-keys\") pod \"keystone-bootstrap-274hf\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " pod="openstack/keystone-bootstrap-274hf" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.460673 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff441522-39b1-47e3-be7d-d590baec8889-scripts\") pod \"horizon-55fff446b9-mbvp9\" (UID: \"ff441522-39b1-47e3-be7d-d590baec8889\") " pod="openstack/horizon-55fff446b9-mbvp9" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.460708 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ff441522-39b1-47e3-be7d-d590baec8889-horizon-secret-key\") pod \"horizon-55fff446b9-mbvp9\" (UID: \"ff441522-39b1-47e3-be7d-d590baec8889\") " pod="openstack/horizon-55fff446b9-mbvp9" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.460736 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff441522-39b1-47e3-be7d-d590baec8889-logs\") pod \"horizon-55fff446b9-mbvp9\" (UID: \"ff441522-39b1-47e3-be7d-d590baec8889\") " pod="openstack/horizon-55fff446b9-mbvp9" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.460830 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xm8mg\" (UniqueName: \"kubernetes.io/projected/ff441522-39b1-47e3-be7d-d590baec8889-kube-api-access-xm8mg\") pod \"horizon-55fff446b9-mbvp9\" (UID: \"ff441522-39b1-47e3-be7d-d590baec8889\") " pod="openstack/horizon-55fff446b9-mbvp9" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.460852 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5ldj\" (UniqueName: \"kubernetes.io/projected/ae433ea4-7305-4d95-96b5-da695d0ed504-kube-api-access-t5ldj\") pod \"keystone-bootstrap-274hf\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " pod="openstack/keystone-bootstrap-274hf" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.460900 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-config-data\") pod \"keystone-bootstrap-274hf\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " pod="openstack/keystone-bootstrap-274hf" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.471583 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-combined-ca-bundle\") pod \"keystone-bootstrap-274hf\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " pod="openstack/keystone-bootstrap-274hf" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.474045 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-config-data\") pod \"keystone-bootstrap-274hf\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " pod="openstack/keystone-bootstrap-274hf" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.480033 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-scripts\") pod \"keystone-bootstrap-274hf\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " pod="openstack/keystone-bootstrap-274hf" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.490265 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-fernet-keys\") pod \"keystone-bootstrap-274hf\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " pod="openstack/keystone-bootstrap-274hf" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.492631 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66dd8c6975-c6tff" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.492951 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-credential-keys\") pod \"keystone-bootstrap-274hf\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " pod="openstack/keystone-bootstrap-274hf" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.514846 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.517227 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.518410 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.538824 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.539082 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.566520 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5ldj\" (UniqueName: \"kubernetes.io/projected/ae433ea4-7305-4d95-96b5-da695d0ed504-kube-api-access-t5ldj\") pod \"keystone-bootstrap-274hf\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " pod="openstack/keystone-bootstrap-274hf" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.574992 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff441522-39b1-47e3-be7d-d590baec8889-scripts\") pod \"horizon-55fff446b9-mbvp9\" (UID: \"ff441522-39b1-47e3-be7d-d590baec8889\") " pod="openstack/horizon-55fff446b9-mbvp9" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.575068 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ff441522-39b1-47e3-be7d-d590baec8889-horizon-secret-key\") pod \"horizon-55fff446b9-mbvp9\" (UID: \"ff441522-39b1-47e3-be7d-d590baec8889\") " pod="openstack/horizon-55fff446b9-mbvp9" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.575102 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff441522-39b1-47e3-be7d-d590baec8889-logs\") pod \"horizon-55fff446b9-mbvp9\" (UID: \"ff441522-39b1-47e3-be7d-d590baec8889\") " pod="openstack/horizon-55fff446b9-mbvp9" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.575161 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xm8mg\" (UniqueName: \"kubernetes.io/projected/ff441522-39b1-47e3-be7d-d590baec8889-kube-api-access-xm8mg\") pod \"horizon-55fff446b9-mbvp9\" (UID: \"ff441522-39b1-47e3-be7d-d590baec8889\") " pod="openstack/horizon-55fff446b9-mbvp9" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.575229 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff441522-39b1-47e3-be7d-d590baec8889-config-data\") pod \"horizon-55fff446b9-mbvp9\" (UID: \"ff441522-39b1-47e3-be7d-d590baec8889\") " pod="openstack/horizon-55fff446b9-mbvp9" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.578138 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff441522-39b1-47e3-be7d-d590baec8889-logs\") pod \"horizon-55fff446b9-mbvp9\" (UID: \"ff441522-39b1-47e3-be7d-d590baec8889\") " pod="openstack/horizon-55fff446b9-mbvp9" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.578896 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff441522-39b1-47e3-be7d-d590baec8889-scripts\") pod \"horizon-55fff446b9-mbvp9\" (UID: \"ff441522-39b1-47e3-be7d-d590baec8889\") " pod="openstack/horizon-55fff446b9-mbvp9" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.576631 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff441522-39b1-47e3-be7d-d590baec8889-config-data\") pod \"horizon-55fff446b9-mbvp9\" (UID: \"ff441522-39b1-47e3-be7d-d590baec8889\") " pod="openstack/horizon-55fff446b9-mbvp9" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.584160 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ff441522-39b1-47e3-be7d-d590baec8889-horizon-secret-key\") pod \"horizon-55fff446b9-mbvp9\" (UID: \"ff441522-39b1-47e3-be7d-d590baec8889\") " pod="openstack/horizon-55fff446b9-mbvp9" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.584237 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-lfx27"] Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.585293 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lfx27" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.592544 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-8s7s9" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.595372 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.597426 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.613682 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-lfx27"] Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.617405 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xm8mg\" (UniqueName: \"kubernetes.io/projected/ff441522-39b1-47e3-be7d-d590baec8889-kube-api-access-xm8mg\") pod \"horizon-55fff446b9-mbvp9\" (UID: \"ff441522-39b1-47e3-be7d-d590baec8889\") " pod="openstack/horizon-55fff446b9-mbvp9" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.636314 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6ff8f88b5f-gnrzd"] Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.637601 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6ff8f88b5f-gnrzd" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.654678 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-nbnz8"] Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.656730 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-nbnz8" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.678792 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " pod="openstack/ceilometer-0" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.678850 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bab95882-60a4-4e02-b3b7-b2a6ae55315f-log-httpd\") pod \"ceilometer-0\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " pod="openstack/ceilometer-0" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.678874 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bab95882-60a4-4e02-b3b7-b2a6ae55315f-run-httpd\") pod \"ceilometer-0\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " pod="openstack/ceilometer-0" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.678904 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8nqvg\" (UniqueName: \"kubernetes.io/projected/bab95882-60a4-4e02-b3b7-b2a6ae55315f-kube-api-access-8nqvg\") pod \"ceilometer-0\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " pod="openstack/ceilometer-0" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.678921 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " pod="openstack/ceilometer-0" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.678936 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-scripts\") pod \"ceilometer-0\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " pod="openstack/ceilometer-0" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.678960 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-config-data\") pod \"ceilometer-0\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " pod="openstack/ceilometer-0" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.684308 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55fff446b9-mbvp9" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.692687 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6ff8f88b5f-gnrzd"] Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.712780 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.713007 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-ktrhh" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.713094 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.734487 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-nbnz8"] Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.778426 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-s4wr5"] Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.779822 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-s4wr5" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.781240 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-scripts\") pod \"cinder-db-sync-nbnz8\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " pod="openstack/cinder-db-sync-nbnz8" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.781273 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bab95882-60a4-4e02-b3b7-b2a6ae55315f-log-httpd\") pod \"ceilometer-0\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " pod="openstack/ceilometer-0" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.781295 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ad9eb8c-d035-456d-a345-5a4da859fc7c-scripts\") pod \"horizon-6ff8f88b5f-gnrzd\" (UID: \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\") " pod="openstack/horizon-6ff8f88b5f-gnrzd" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.781316 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bab95882-60a4-4e02-b3b7-b2a6ae55315f-run-httpd\") pod \"ceilometer-0\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " pod="openstack/ceilometer-0" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.781345 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wwj2\" (UniqueName: \"kubernetes.io/projected/ce9f3ab9-097d-4d0d-8082-08019bce834c-kube-api-access-6wwj2\") pod \"cinder-db-sync-nbnz8\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " pod="openstack/cinder-db-sync-nbnz8" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.781361 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8nqvg\" (UniqueName: \"kubernetes.io/projected/bab95882-60a4-4e02-b3b7-b2a6ae55315f-kube-api-access-8nqvg\") pod \"ceilometer-0\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " pod="openstack/ceilometer-0" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.781380 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " pod="openstack/ceilometer-0" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.781396 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-scripts\") pod \"ceilometer-0\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " pod="openstack/ceilometer-0" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.781412 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ad9eb8c-d035-456d-a345-5a4da859fc7c-config-data\") pod \"horizon-6ff8f88b5f-gnrzd\" (UID: \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\") " pod="openstack/horizon-6ff8f88b5f-gnrzd" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.781433 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-config-data\") pod \"ceilometer-0\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " pod="openstack/ceilometer-0" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.781452 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-db-sync-config-data\") pod \"cinder-db-sync-nbnz8\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " pod="openstack/cinder-db-sync-nbnz8" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.781474 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggwmj\" (UniqueName: \"kubernetes.io/projected/b48430a2-ef1d-4747-97a4-4b6933fcea4f-kube-api-access-ggwmj\") pod \"neutron-db-sync-lfx27\" (UID: \"b48430a2-ef1d-4747-97a4-4b6933fcea4f\") " pod="openstack/neutron-db-sync-lfx27" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.781491 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-config-data\") pod \"cinder-db-sync-nbnz8\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " pod="openstack/cinder-db-sync-nbnz8" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.781515 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1ad9eb8c-d035-456d-a345-5a4da859fc7c-horizon-secret-key\") pod \"horizon-6ff8f88b5f-gnrzd\" (UID: \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\") " pod="openstack/horizon-6ff8f88b5f-gnrzd" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.781528 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ad9eb8c-d035-456d-a345-5a4da859fc7c-logs\") pod \"horizon-6ff8f88b5f-gnrzd\" (UID: \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\") " pod="openstack/horizon-6ff8f88b5f-gnrzd" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.781542 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-combined-ca-bundle\") pod \"cinder-db-sync-nbnz8\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " pod="openstack/cinder-db-sync-nbnz8" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.781572 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvtfd\" (UniqueName: \"kubernetes.io/projected/1ad9eb8c-d035-456d-a345-5a4da859fc7c-kube-api-access-bvtfd\") pod \"horizon-6ff8f88b5f-gnrzd\" (UID: \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\") " pod="openstack/horizon-6ff8f88b5f-gnrzd" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.781594 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b48430a2-ef1d-4747-97a4-4b6933fcea4f-combined-ca-bundle\") pod \"neutron-db-sync-lfx27\" (UID: \"b48430a2-ef1d-4747-97a4-4b6933fcea4f\") " pod="openstack/neutron-db-sync-lfx27" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.781621 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b48430a2-ef1d-4747-97a4-4b6933fcea4f-config\") pod \"neutron-db-sync-lfx27\" (UID: \"b48430a2-ef1d-4747-97a4-4b6933fcea4f\") " pod="openstack/neutron-db-sync-lfx27" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.781654 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ce9f3ab9-097d-4d0d-8082-08019bce834c-etc-machine-id\") pod \"cinder-db-sync-nbnz8\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " pod="openstack/cinder-db-sync-nbnz8" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.781687 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " pod="openstack/ceilometer-0" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.782816 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-x8k6m" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.786531 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.786591 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " pod="openstack/ceilometer-0" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.786894 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bab95882-60a4-4e02-b3b7-b2a6ae55315f-log-httpd\") pod \"ceilometer-0\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " pod="openstack/ceilometer-0" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.789151 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bab95882-60a4-4e02-b3b7-b2a6ae55315f-run-httpd\") pod \"ceilometer-0\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " pod="openstack/ceilometer-0" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.790568 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-scripts\") pod \"ceilometer-0\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " pod="openstack/ceilometer-0" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.794587 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " pod="openstack/ceilometer-0" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.795659 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-config-data\") pod \"ceilometer-0\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " pod="openstack/ceilometer-0" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.803093 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-s4wr5"] Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.815173 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8nqvg\" (UniqueName: \"kubernetes.io/projected/bab95882-60a4-4e02-b3b7-b2a6ae55315f-kube-api-access-8nqvg\") pod \"ceilometer-0\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " pod="openstack/ceilometer-0" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.828767 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66dd8c6975-c6tff"] Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.841242 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-jr2lq"] Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.842929 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jr2lq" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.849563 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-274hf" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.849737 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-bjd4l" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.855935 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-694dbb6647-fzs9q"] Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.856225 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.882831 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.887706 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ad9eb8c-d035-456d-a345-5a4da859fc7c-config-data\") pod \"horizon-6ff8f88b5f-gnrzd\" (UID: \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\") " pod="openstack/horizon-6ff8f88b5f-gnrzd" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.887791 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-db-sync-config-data\") pod \"cinder-db-sync-nbnz8\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " pod="openstack/cinder-db-sync-nbnz8" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.887842 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggwmj\" (UniqueName: \"kubernetes.io/projected/b48430a2-ef1d-4747-97a4-4b6933fcea4f-kube-api-access-ggwmj\") pod \"neutron-db-sync-lfx27\" (UID: \"b48430a2-ef1d-4747-97a4-4b6933fcea4f\") " pod="openstack/neutron-db-sync-lfx27" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.887877 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-config-data\") pod \"cinder-db-sync-nbnz8\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " pod="openstack/cinder-db-sync-nbnz8" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.887921 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-combined-ca-bundle\") pod \"cinder-db-sync-nbnz8\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " pod="openstack/cinder-db-sync-nbnz8" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.887947 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1ad9eb8c-d035-456d-a345-5a4da859fc7c-horizon-secret-key\") pod \"horizon-6ff8f88b5f-gnrzd\" (UID: \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\") " pod="openstack/horizon-6ff8f88b5f-gnrzd" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.887977 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ad9eb8c-d035-456d-a345-5a4da859fc7c-logs\") pod \"horizon-6ff8f88b5f-gnrzd\" (UID: \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\") " pod="openstack/horizon-6ff8f88b5f-gnrzd" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.888036 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvtfd\" (UniqueName: \"kubernetes.io/projected/1ad9eb8c-d035-456d-a345-5a4da859fc7c-kube-api-access-bvtfd\") pod \"horizon-6ff8f88b5f-gnrzd\" (UID: \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\") " pod="openstack/horizon-6ff8f88b5f-gnrzd" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.888071 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b48430a2-ef1d-4747-97a4-4b6933fcea4f-combined-ca-bundle\") pod \"neutron-db-sync-lfx27\" (UID: \"b48430a2-ef1d-4747-97a4-4b6933fcea4f\") " pod="openstack/neutron-db-sync-lfx27" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.888109 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29fb637-ba21-4cf9-ae4d-cd9720dd5af4-combined-ca-bundle\") pod \"barbican-db-sync-s4wr5\" (UID: \"d29fb637-ba21-4cf9-ae4d-cd9720dd5af4\") " pod="openstack/barbican-db-sync-s4wr5" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.888156 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b48430a2-ef1d-4747-97a4-4b6933fcea4f-config\") pod \"neutron-db-sync-lfx27\" (UID: \"b48430a2-ef1d-4747-97a4-4b6933fcea4f\") " pod="openstack/neutron-db-sync-lfx27" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.888192 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ce9f3ab9-097d-4d0d-8082-08019bce834c-etc-machine-id\") pod \"cinder-db-sync-nbnz8\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " pod="openstack/cinder-db-sync-nbnz8" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.888286 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-scripts\") pod \"cinder-db-sync-nbnz8\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " pod="openstack/cinder-db-sync-nbnz8" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.888337 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ad9eb8c-d035-456d-a345-5a4da859fc7c-scripts\") pod \"horizon-6ff8f88b5f-gnrzd\" (UID: \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\") " pod="openstack/horizon-6ff8f88b5f-gnrzd" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.888381 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d29fb637-ba21-4cf9-ae4d-cd9720dd5af4-db-sync-config-data\") pod \"barbican-db-sync-s4wr5\" (UID: \"d29fb637-ba21-4cf9-ae4d-cd9720dd5af4\") " pod="openstack/barbican-db-sync-s4wr5" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.888418 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b4rm6\" (UniqueName: \"kubernetes.io/projected/d29fb637-ba21-4cf9-ae4d-cd9720dd5af4-kube-api-access-b4rm6\") pod \"barbican-db-sync-s4wr5\" (UID: \"d29fb637-ba21-4cf9-ae4d-cd9720dd5af4\") " pod="openstack/barbican-db-sync-s4wr5" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.888466 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wwj2\" (UniqueName: \"kubernetes.io/projected/ce9f3ab9-097d-4d0d-8082-08019bce834c-kube-api-access-6wwj2\") pod \"cinder-db-sync-nbnz8\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " pod="openstack/cinder-db-sync-nbnz8" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.890456 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ce9f3ab9-097d-4d0d-8082-08019bce834c-etc-machine-id\") pod \"cinder-db-sync-nbnz8\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " pod="openstack/cinder-db-sync-nbnz8" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.903046 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ad9eb8c-d035-456d-a345-5a4da859fc7c-logs\") pod \"horizon-6ff8f88b5f-gnrzd\" (UID: \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\") " pod="openstack/horizon-6ff8f88b5f-gnrzd" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.910793 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ad9eb8c-d035-456d-a345-5a4da859fc7c-scripts\") pod \"horizon-6ff8f88b5f-gnrzd\" (UID: \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\") " pod="openstack/horizon-6ff8f88b5f-gnrzd" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.912815 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.914515 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ad9eb8c-d035-456d-a345-5a4da859fc7c-config-data\") pod \"horizon-6ff8f88b5f-gnrzd\" (UID: \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\") " pod="openstack/horizon-6ff8f88b5f-gnrzd" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.916585 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/b48430a2-ef1d-4747-97a4-4b6933fcea4f-config\") pod \"neutron-db-sync-lfx27\" (UID: \"b48430a2-ef1d-4747-97a4-4b6933fcea4f\") " pod="openstack/neutron-db-sync-lfx27" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.931604 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b48430a2-ef1d-4747-97a4-4b6933fcea4f-combined-ca-bundle\") pod \"neutron-db-sync-lfx27\" (UID: \"b48430a2-ef1d-4747-97a4-4b6933fcea4f\") " pod="openstack/neutron-db-sync-lfx27" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.930270 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-scripts\") pod \"cinder-db-sync-nbnz8\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " pod="openstack/cinder-db-sync-nbnz8" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.934517 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvtfd\" (UniqueName: \"kubernetes.io/projected/1ad9eb8c-d035-456d-a345-5a4da859fc7c-kube-api-access-bvtfd\") pod \"horizon-6ff8f88b5f-gnrzd\" (UID: \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\") " pod="openstack/horizon-6ff8f88b5f-gnrzd" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.935160 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-combined-ca-bundle\") pod \"cinder-db-sync-nbnz8\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " pod="openstack/cinder-db-sync-nbnz8" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.935286 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-config-data\") pod \"cinder-db-sync-nbnz8\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " pod="openstack/cinder-db-sync-nbnz8" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.935554 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggwmj\" (UniqueName: \"kubernetes.io/projected/b48430a2-ef1d-4747-97a4-4b6933fcea4f-kube-api-access-ggwmj\") pod \"neutron-db-sync-lfx27\" (UID: \"b48430a2-ef1d-4747-97a4-4b6933fcea4f\") " pod="openstack/neutron-db-sync-lfx27" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.937341 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-db-sync-config-data\") pod \"cinder-db-sync-nbnz8\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " pod="openstack/cinder-db-sync-nbnz8" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.937857 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-694dbb6647-fzs9q"] Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.946392 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wwj2\" (UniqueName: \"kubernetes.io/projected/ce9f3ab9-097d-4d0d-8082-08019bce834c-kube-api-access-6wwj2\") pod \"cinder-db-sync-nbnz8\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " pod="openstack/cinder-db-sync-nbnz8" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.960447 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1ad9eb8c-d035-456d-a345-5a4da859fc7c-horizon-secret-key\") pod \"horizon-6ff8f88b5f-gnrzd\" (UID: \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\") " pod="openstack/horizon-6ff8f88b5f-gnrzd" Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.965115 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-jr2lq"] Dec 09 15:48:59 crc kubenswrapper[4894]: I1209 15:48:59.966494 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.009794 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-ovsdbserver-sb\") pod \"dnsmasq-dns-694dbb6647-fzs9q\" (UID: \"f20c656d-4752-41ce-a1aa-aa6980430e90\") " pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.010188 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d3e1277-28d7-4aa1-9121-75d859ae4688-logs\") pod \"placement-db-sync-jr2lq\" (UID: \"7d3e1277-28d7-4aa1-9121-75d859ae4688\") " pod="openstack/placement-db-sync-jr2lq" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.010284 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bc5tl\" (UniqueName: \"kubernetes.io/projected/7d3e1277-28d7-4aa1-9121-75d859ae4688-kube-api-access-bc5tl\") pod \"placement-db-sync-jr2lq\" (UID: \"7d3e1277-28d7-4aa1-9121-75d859ae4688\") " pod="openstack/placement-db-sync-jr2lq" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.010493 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d29fb637-ba21-4cf9-ae4d-cd9720dd5af4-db-sync-config-data\") pod \"barbican-db-sync-s4wr5\" (UID: \"d29fb637-ba21-4cf9-ae4d-cd9720dd5af4\") " pod="openstack/barbican-db-sync-s4wr5" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.010540 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b4rm6\" (UniqueName: \"kubernetes.io/projected/d29fb637-ba21-4cf9-ae4d-cd9720dd5af4-kube-api-access-b4rm6\") pod \"barbican-db-sync-s4wr5\" (UID: \"d29fb637-ba21-4cf9-ae4d-cd9720dd5af4\") " pod="openstack/barbican-db-sync-s4wr5" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.010631 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc5nm\" (UniqueName: \"kubernetes.io/projected/f20c656d-4752-41ce-a1aa-aa6980430e90-kube-api-access-qc5nm\") pod \"dnsmasq-dns-694dbb6647-fzs9q\" (UID: \"f20c656d-4752-41ce-a1aa-aa6980430e90\") " pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.015686 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d3e1277-28d7-4aa1-9121-75d859ae4688-config-data\") pod \"placement-db-sync-jr2lq\" (UID: \"7d3e1277-28d7-4aa1-9121-75d859ae4688\") " pod="openstack/placement-db-sync-jr2lq" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.015790 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-config\") pod \"dnsmasq-dns-694dbb6647-fzs9q\" (UID: \"f20c656d-4752-41ce-a1aa-aa6980430e90\") " pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.015813 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d3e1277-28d7-4aa1-9121-75d859ae4688-combined-ca-bundle\") pod \"placement-db-sync-jr2lq\" (UID: \"7d3e1277-28d7-4aa1-9121-75d859ae4688\") " pod="openstack/placement-db-sync-jr2lq" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.015934 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29fb637-ba21-4cf9-ae4d-cd9720dd5af4-combined-ca-bundle\") pod \"barbican-db-sync-s4wr5\" (UID: \"d29fb637-ba21-4cf9-ae4d-cd9720dd5af4\") " pod="openstack/barbican-db-sync-s4wr5" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.016075 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-ovsdbserver-nb\") pod \"dnsmasq-dns-694dbb6647-fzs9q\" (UID: \"f20c656d-4752-41ce-a1aa-aa6980430e90\") " pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.016183 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-dns-svc\") pod \"dnsmasq-dns-694dbb6647-fzs9q\" (UID: \"f20c656d-4752-41ce-a1aa-aa6980430e90\") " pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.016221 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d3e1277-28d7-4aa1-9121-75d859ae4688-scripts\") pod \"placement-db-sync-jr2lq\" (UID: \"7d3e1277-28d7-4aa1-9121-75d859ae4688\") " pod="openstack/placement-db-sync-jr2lq" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.022975 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d29fb637-ba21-4cf9-ae4d-cd9720dd5af4-db-sync-config-data\") pod \"barbican-db-sync-s4wr5\" (UID: \"d29fb637-ba21-4cf9-ae4d-cd9720dd5af4\") " pod="openstack/barbican-db-sync-s4wr5" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.030710 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29fb637-ba21-4cf9-ae4d-cd9720dd5af4-combined-ca-bundle\") pod \"barbican-db-sync-s4wr5\" (UID: \"d29fb637-ba21-4cf9-ae4d-cd9720dd5af4\") " pod="openstack/barbican-db-sync-s4wr5" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.043102 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b4rm6\" (UniqueName: \"kubernetes.io/projected/d29fb637-ba21-4cf9-ae4d-cd9720dd5af4-kube-api-access-b4rm6\") pod \"barbican-db-sync-s4wr5\" (UID: \"d29fb637-ba21-4cf9-ae4d-cd9720dd5af4\") " pod="openstack/barbican-db-sync-s4wr5" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.114297 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lfx27" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.124414 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-ovsdbserver-sb\") pod \"dnsmasq-dns-694dbb6647-fzs9q\" (UID: \"f20c656d-4752-41ce-a1aa-aa6980430e90\") " pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.124457 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d3e1277-28d7-4aa1-9121-75d859ae4688-logs\") pod \"placement-db-sync-jr2lq\" (UID: \"7d3e1277-28d7-4aa1-9121-75d859ae4688\") " pod="openstack/placement-db-sync-jr2lq" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.124498 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bc5tl\" (UniqueName: \"kubernetes.io/projected/7d3e1277-28d7-4aa1-9121-75d859ae4688-kube-api-access-bc5tl\") pod \"placement-db-sync-jr2lq\" (UID: \"7d3e1277-28d7-4aa1-9121-75d859ae4688\") " pod="openstack/placement-db-sync-jr2lq" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.124571 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc5nm\" (UniqueName: \"kubernetes.io/projected/f20c656d-4752-41ce-a1aa-aa6980430e90-kube-api-access-qc5nm\") pod \"dnsmasq-dns-694dbb6647-fzs9q\" (UID: \"f20c656d-4752-41ce-a1aa-aa6980430e90\") " pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.124604 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d3e1277-28d7-4aa1-9121-75d859ae4688-config-data\") pod \"placement-db-sync-jr2lq\" (UID: \"7d3e1277-28d7-4aa1-9121-75d859ae4688\") " pod="openstack/placement-db-sync-jr2lq" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.124681 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-config\") pod \"dnsmasq-dns-694dbb6647-fzs9q\" (UID: \"f20c656d-4752-41ce-a1aa-aa6980430e90\") " pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.124702 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d3e1277-28d7-4aa1-9121-75d859ae4688-combined-ca-bundle\") pod \"placement-db-sync-jr2lq\" (UID: \"7d3e1277-28d7-4aa1-9121-75d859ae4688\") " pod="openstack/placement-db-sync-jr2lq" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.124760 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-ovsdbserver-nb\") pod \"dnsmasq-dns-694dbb6647-fzs9q\" (UID: \"f20c656d-4752-41ce-a1aa-aa6980430e90\") " pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.124805 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-dns-svc\") pod \"dnsmasq-dns-694dbb6647-fzs9q\" (UID: \"f20c656d-4752-41ce-a1aa-aa6980430e90\") " pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.124832 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d3e1277-28d7-4aa1-9121-75d859ae4688-scripts\") pod \"placement-db-sync-jr2lq\" (UID: \"7d3e1277-28d7-4aa1-9121-75d859ae4688\") " pod="openstack/placement-db-sync-jr2lq" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.127816 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d3e1277-28d7-4aa1-9121-75d859ae4688-logs\") pod \"placement-db-sync-jr2lq\" (UID: \"7d3e1277-28d7-4aa1-9121-75d859ae4688\") " pod="openstack/placement-db-sync-jr2lq" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.128059 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-ovsdbserver-sb\") pod \"dnsmasq-dns-694dbb6647-fzs9q\" (UID: \"f20c656d-4752-41ce-a1aa-aa6980430e90\") " pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.129262 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-config\") pod \"dnsmasq-dns-694dbb6647-fzs9q\" (UID: \"f20c656d-4752-41ce-a1aa-aa6980430e90\") " pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.129307 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-ovsdbserver-nb\") pod \"dnsmasq-dns-694dbb6647-fzs9q\" (UID: \"f20c656d-4752-41ce-a1aa-aa6980430e90\") " pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.130987 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-dns-svc\") pod \"dnsmasq-dns-694dbb6647-fzs9q\" (UID: \"f20c656d-4752-41ce-a1aa-aa6980430e90\") " pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.132660 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d3e1277-28d7-4aa1-9121-75d859ae4688-scripts\") pod \"placement-db-sync-jr2lq\" (UID: \"7d3e1277-28d7-4aa1-9121-75d859ae4688\") " pod="openstack/placement-db-sync-jr2lq" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.134086 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d3e1277-28d7-4aa1-9121-75d859ae4688-combined-ca-bundle\") pod \"placement-db-sync-jr2lq\" (UID: \"7d3e1277-28d7-4aa1-9121-75d859ae4688\") " pod="openstack/placement-db-sync-jr2lq" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.135959 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6ff8f88b5f-gnrzd" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.143344 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d3e1277-28d7-4aa1-9121-75d859ae4688-config-data\") pod \"placement-db-sync-jr2lq\" (UID: \"7d3e1277-28d7-4aa1-9121-75d859ae4688\") " pod="openstack/placement-db-sync-jr2lq" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.155976 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bc5tl\" (UniqueName: \"kubernetes.io/projected/7d3e1277-28d7-4aa1-9121-75d859ae4688-kube-api-access-bc5tl\") pod \"placement-db-sync-jr2lq\" (UID: \"7d3e1277-28d7-4aa1-9121-75d859ae4688\") " pod="openstack/placement-db-sync-jr2lq" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.162779 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc5nm\" (UniqueName: \"kubernetes.io/projected/f20c656d-4752-41ce-a1aa-aa6980430e90-kube-api-access-qc5nm\") pod \"dnsmasq-dns-694dbb6647-fzs9q\" (UID: \"f20c656d-4752-41ce-a1aa-aa6980430e90\") " pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.189649 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-nbnz8" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.196768 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66dd8c6975-c6tff"] Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.273444 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-s4wr5" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.283565 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jr2lq" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.314623 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.332885 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-55fff446b9-mbvp9"] Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.371609 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:49:00 crc kubenswrapper[4894]: W1209 15:49:00.412260 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podbab95882_60a4_4e02_b3b7_b2a6ae55315f.slice/crio-7ff237304503b28262d2010661c11d69f410f1f9f57bdf215964ba305c37ef41 WatchSource:0}: Error finding container 7ff237304503b28262d2010661c11d69f410f1f9f57bdf215964ba305c37ef41: Status 404 returned error can't find the container with id 7ff237304503b28262d2010661c11d69f410f1f9f57bdf215964ba305c37ef41 Dec 09 15:49:00 crc kubenswrapper[4894]: W1209 15:49:00.428630 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podff441522_39b1_47e3_be7d_d590baec8889.slice/crio-2f113461bf6f89fb201c28e5cdcd43074cd02b5a622451af9af350a6df30d448 WatchSource:0}: Error finding container 2f113461bf6f89fb201c28e5cdcd43074cd02b5a622451af9af350a6df30d448: Status 404 returned error can't find the container with id 2f113461bf6f89fb201c28e5cdcd43074cd02b5a622451af9af350a6df30d448 Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.504146 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-274hf"] Dec 09 15:49:00 crc kubenswrapper[4894]: I1209 15:49:00.760708 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-lfx27"] Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:00.944188 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6ff8f88b5f-gnrzd"] Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:00.979115 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-nbnz8"] Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:00.989620 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55fff446b9-mbvp9" event={"ID":"ff441522-39b1-47e3-be7d-d590baec8889","Type":"ContainerStarted","Data":"2f113461bf6f89fb201c28e5cdcd43074cd02b5a622451af9af350a6df30d448"} Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.002672 4894 generic.go:334] "Generic (PLEG): container finished" podID="453feb9e-4bbd-49b8-9719-07f97af54b46" containerID="166be08764e066427ee921580f9db9c6eeb23b7a34359ff64e1dd0a93ae54ae2" exitCode=0 Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.002729 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66dd8c6975-c6tff" event={"ID":"453feb9e-4bbd-49b8-9719-07f97af54b46","Type":"ContainerDied","Data":"166be08764e066427ee921580f9db9c6eeb23b7a34359ff64e1dd0a93ae54ae2"} Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.002753 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66dd8c6975-c6tff" event={"ID":"453feb9e-4bbd-49b8-9719-07f97af54b46","Type":"ContainerStarted","Data":"357d71490f25db3c3df0259eefc713c4d4afc1148392030d8825cc8b1985aa8e"} Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.034987 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bab95882-60a4-4e02-b3b7-b2a6ae55315f","Type":"ContainerStarted","Data":"7ff237304503b28262d2010661c11d69f410f1f9f57bdf215964ba305c37ef41"} Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.055937 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lfx27" event={"ID":"b48430a2-ef1d-4747-97a4-4b6933fcea4f","Type":"ContainerStarted","Data":"a53ee65412c8d5fe99865aaf2cfd66727c78880d5f6b10e6d1e5f2045a2f25e1"} Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.062707 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-274hf" event={"ID":"ae433ea4-7305-4d95-96b5-da695d0ed504","Type":"ContainerStarted","Data":"466350b715f0a001b96f8b05271b4b5a567f93e5f14fc6dd76ac361dc90e20d6"} Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.062757 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-274hf" event={"ID":"ae433ea4-7305-4d95-96b5-da695d0ed504","Type":"ContainerStarted","Data":"8bbf03beb82f7a65c3c169b766814f6fd1c6d072354bd302596626387be7bb82"} Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.107299 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-lfx27" podStartSLOduration=2.107275456 podStartE2EDuration="2.107275456s" podCreationTimestamp="2025-12-09 15:48:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:49:01.078203777 +0000 UTC m=+1035.397414446" watchObservedRunningTime="2025-12-09 15:49:01.107275456 +0000 UTC m=+1035.426486125" Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.130706 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-s4wr5"] Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.134910 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-274hf" podStartSLOduration=2.134888308 podStartE2EDuration="2.134888308s" podCreationTimestamp="2025-12-09 15:48:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:49:01.111440445 +0000 UTC m=+1035.430651114" watchObservedRunningTime="2025-12-09 15:49:01.134888308 +0000 UTC m=+1035.454098977" Dec 09 15:49:01 crc kubenswrapper[4894]: W1209 15:49:01.139478 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf20c656d_4752_41ce_a1aa_aa6980430e90.slice/crio-0690d6aa19b142e8028f2fb342968df9a4acb8bcbb77a1d1c891166839c8b990 WatchSource:0}: Error finding container 0690d6aa19b142e8028f2fb342968df9a4acb8bcbb77a1d1c891166839c8b990: Status 404 returned error can't find the container with id 0690d6aa19b142e8028f2fb342968df9a4acb8bcbb77a1d1c891166839c8b990 Dec 09 15:49:01 crc kubenswrapper[4894]: W1209 15:49:01.145959 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd29fb637_ba21_4cf9_ae4d_cd9720dd5af4.slice/crio-f532e366948e0e6b1b702f291b0c0e8660fce048f415d2b7215f7a995067200c WatchSource:0}: Error finding container f532e366948e0e6b1b702f291b0c0e8660fce048f415d2b7215f7a995067200c: Status 404 returned error can't find the container with id f532e366948e0e6b1b702f291b0c0e8660fce048f415d2b7215f7a995067200c Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.150114 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-694dbb6647-fzs9q"] Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.203362 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-jr2lq"] Dec 09 15:49:01 crc kubenswrapper[4894]: W1209 15:49:01.232048 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7d3e1277_28d7_4aa1_9121_75d859ae4688.slice/crio-79c97c797089c4b3b8dc88a13bacdb638d7e52432f17a350ea19b191336ead57 WatchSource:0}: Error finding container 79c97c797089c4b3b8dc88a13bacdb638d7e52432f17a350ea19b191336ead57: Status 404 returned error can't find the container with id 79c97c797089c4b3b8dc88a13bacdb638d7e52432f17a350ea19b191336ead57 Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.520041 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-55fff446b9-mbvp9"] Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.569701 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-649c9bf67-kcmj8"] Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.571421 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-649c9bf67-kcmj8" Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.627341 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-649c9bf67-kcmj8"] Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.654971 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18cc81c1-7bef-41c2-89b7-e507624504f4-config-data\") pod \"horizon-649c9bf67-kcmj8\" (UID: \"18cc81c1-7bef-41c2-89b7-e507624504f4\") " pod="openstack/horizon-649c9bf67-kcmj8" Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.655050 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18cc81c1-7bef-41c2-89b7-e507624504f4-scripts\") pod \"horizon-649c9bf67-kcmj8\" (UID: \"18cc81c1-7bef-41c2-89b7-e507624504f4\") " pod="openstack/horizon-649c9bf67-kcmj8" Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.655074 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/18cc81c1-7bef-41c2-89b7-e507624504f4-horizon-secret-key\") pod \"horizon-649c9bf67-kcmj8\" (UID: \"18cc81c1-7bef-41c2-89b7-e507624504f4\") " pod="openstack/horizon-649c9bf67-kcmj8" Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.655124 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq92x\" (UniqueName: \"kubernetes.io/projected/18cc81c1-7bef-41c2-89b7-e507624504f4-kube-api-access-pq92x\") pod \"horizon-649c9bf67-kcmj8\" (UID: \"18cc81c1-7bef-41c2-89b7-e507624504f4\") " pod="openstack/horizon-649c9bf67-kcmj8" Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.655233 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18cc81c1-7bef-41c2-89b7-e507624504f4-logs\") pod \"horizon-649c9bf67-kcmj8\" (UID: \"18cc81c1-7bef-41c2-89b7-e507624504f4\") " pod="openstack/horizon-649c9bf67-kcmj8" Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.758743 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18cc81c1-7bef-41c2-89b7-e507624504f4-logs\") pod \"horizon-649c9bf67-kcmj8\" (UID: \"18cc81c1-7bef-41c2-89b7-e507624504f4\") " pod="openstack/horizon-649c9bf67-kcmj8" Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.758897 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18cc81c1-7bef-41c2-89b7-e507624504f4-config-data\") pod \"horizon-649c9bf67-kcmj8\" (UID: \"18cc81c1-7bef-41c2-89b7-e507624504f4\") " pod="openstack/horizon-649c9bf67-kcmj8" Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.758942 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18cc81c1-7bef-41c2-89b7-e507624504f4-scripts\") pod \"horizon-649c9bf67-kcmj8\" (UID: \"18cc81c1-7bef-41c2-89b7-e507624504f4\") " pod="openstack/horizon-649c9bf67-kcmj8" Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.758984 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/18cc81c1-7bef-41c2-89b7-e507624504f4-horizon-secret-key\") pod \"horizon-649c9bf67-kcmj8\" (UID: \"18cc81c1-7bef-41c2-89b7-e507624504f4\") " pod="openstack/horizon-649c9bf67-kcmj8" Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.759016 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq92x\" (UniqueName: \"kubernetes.io/projected/18cc81c1-7bef-41c2-89b7-e507624504f4-kube-api-access-pq92x\") pod \"horizon-649c9bf67-kcmj8\" (UID: \"18cc81c1-7bef-41c2-89b7-e507624504f4\") " pod="openstack/horizon-649c9bf67-kcmj8" Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.759810 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18cc81c1-7bef-41c2-89b7-e507624504f4-logs\") pod \"horizon-649c9bf67-kcmj8\" (UID: \"18cc81c1-7bef-41c2-89b7-e507624504f4\") " pod="openstack/horizon-649c9bf67-kcmj8" Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.761143 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18cc81c1-7bef-41c2-89b7-e507624504f4-config-data\") pod \"horizon-649c9bf67-kcmj8\" (UID: \"18cc81c1-7bef-41c2-89b7-e507624504f4\") " pod="openstack/horizon-649c9bf67-kcmj8" Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.763794 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18cc81c1-7bef-41c2-89b7-e507624504f4-scripts\") pod \"horizon-649c9bf67-kcmj8\" (UID: \"18cc81c1-7bef-41c2-89b7-e507624504f4\") " pod="openstack/horizon-649c9bf67-kcmj8" Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.768476 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/18cc81c1-7bef-41c2-89b7-e507624504f4-horizon-secret-key\") pod \"horizon-649c9bf67-kcmj8\" (UID: \"18cc81c1-7bef-41c2-89b7-e507624504f4\") " pod="openstack/horizon-649c9bf67-kcmj8" Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.775291 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq92x\" (UniqueName: \"kubernetes.io/projected/18cc81c1-7bef-41c2-89b7-e507624504f4-kube-api-access-pq92x\") pod \"horizon-649c9bf67-kcmj8\" (UID: \"18cc81c1-7bef-41c2-89b7-e507624504f4\") " pod="openstack/horizon-649c9bf67-kcmj8" Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.928228 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:49:01 crc kubenswrapper[4894]: I1209 15:49:01.962612 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-649c9bf67-kcmj8" Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.085963 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jr2lq" event={"ID":"7d3e1277-28d7-4aa1-9121-75d859ae4688","Type":"ContainerStarted","Data":"79c97c797089c4b3b8dc88a13bacdb638d7e52432f17a350ea19b191336ead57"} Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.167241 4894 generic.go:334] "Generic (PLEG): container finished" podID="f20c656d-4752-41ce-a1aa-aa6980430e90" containerID="9fbc25fcc1a3e6babc8975e83d644bc14327478c515a5f66a069e085fa1776b0" exitCode=0 Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.184867 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" event={"ID":"f20c656d-4752-41ce-a1aa-aa6980430e90","Type":"ContainerDied","Data":"9fbc25fcc1a3e6babc8975e83d644bc14327478c515a5f66a069e085fa1776b0"} Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.185175 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" event={"ID":"f20c656d-4752-41ce-a1aa-aa6980430e90","Type":"ContainerStarted","Data":"0690d6aa19b142e8028f2fb342968df9a4acb8bcbb77a1d1c891166839c8b990"} Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.185189 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-nbnz8" event={"ID":"ce9f3ab9-097d-4d0d-8082-08019bce834c","Type":"ContainerStarted","Data":"16473a4acae940ea360630bcbed55206fc73095e0003fd944f4461f7506d5cff"} Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.185202 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lfx27" event={"ID":"b48430a2-ef1d-4747-97a4-4b6933fcea4f","Type":"ContainerStarted","Data":"b8a1f501c7d2757a2ff9a88f8716626e02ee6c30fae871162ab553918ea821dc"} Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.185838 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66dd8c6975-c6tff" Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.187375 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-s4wr5" event={"ID":"d29fb637-ba21-4cf9-ae4d-cd9720dd5af4","Type":"ContainerStarted","Data":"f532e366948e0e6b1b702f291b0c0e8660fce048f415d2b7215f7a995067200c"} Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.212939 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6ff8f88b5f-gnrzd" event={"ID":"1ad9eb8c-d035-456d-a345-5a4da859fc7c","Type":"ContainerStarted","Data":"66c9c2c13b0b9e90ab27d0f68a9166495a171eaaf587553b4939558bd9bcbbdc"} Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.268047 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-dns-svc\") pod \"453feb9e-4bbd-49b8-9719-07f97af54b46\" (UID: \"453feb9e-4bbd-49b8-9719-07f97af54b46\") " Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.268108 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rcx45\" (UniqueName: \"kubernetes.io/projected/453feb9e-4bbd-49b8-9719-07f97af54b46-kube-api-access-rcx45\") pod \"453feb9e-4bbd-49b8-9719-07f97af54b46\" (UID: \"453feb9e-4bbd-49b8-9719-07f97af54b46\") " Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.268157 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-config\") pod \"453feb9e-4bbd-49b8-9719-07f97af54b46\" (UID: \"453feb9e-4bbd-49b8-9719-07f97af54b46\") " Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.268502 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-ovsdbserver-sb\") pod \"453feb9e-4bbd-49b8-9719-07f97af54b46\" (UID: \"453feb9e-4bbd-49b8-9719-07f97af54b46\") " Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.268584 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-ovsdbserver-nb\") pod \"453feb9e-4bbd-49b8-9719-07f97af54b46\" (UID: \"453feb9e-4bbd-49b8-9719-07f97af54b46\") " Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.273753 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/453feb9e-4bbd-49b8-9719-07f97af54b46-kube-api-access-rcx45" (OuterVolumeSpecName: "kube-api-access-rcx45") pod "453feb9e-4bbd-49b8-9719-07f97af54b46" (UID: "453feb9e-4bbd-49b8-9719-07f97af54b46"). InnerVolumeSpecName "kube-api-access-rcx45". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.290970 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "453feb9e-4bbd-49b8-9719-07f97af54b46" (UID: "453feb9e-4bbd-49b8-9719-07f97af54b46"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.294990 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "453feb9e-4bbd-49b8-9719-07f97af54b46" (UID: "453feb9e-4bbd-49b8-9719-07f97af54b46"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:49:02 crc kubenswrapper[4894]: E1209 15:49:02.309687 4894 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-config podName:453feb9e-4bbd-49b8-9719-07f97af54b46 nodeName:}" failed. No retries permitted until 2025-12-09 15:49:02.809658862 +0000 UTC m=+1037.128869531 (durationBeforeRetry 500ms). Error: error cleaning subPath mounts for volume "config" (UniqueName: "kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-config") pod "453feb9e-4bbd-49b8-9719-07f97af54b46" (UID: "453feb9e-4bbd-49b8-9719-07f97af54b46") : error deleting /var/lib/kubelet/pods/453feb9e-4bbd-49b8-9719-07f97af54b46/volume-subpaths: remove /var/lib/kubelet/pods/453feb9e-4bbd-49b8-9719-07f97af54b46/volume-subpaths: no such file or directory Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.309987 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "453feb9e-4bbd-49b8-9719-07f97af54b46" (UID: "453feb9e-4bbd-49b8-9719-07f97af54b46"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.373394 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rcx45\" (UniqueName: \"kubernetes.io/projected/453feb9e-4bbd-49b8-9719-07f97af54b46-kube-api-access-rcx45\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.373420 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.373429 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.373438 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.537411 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-649c9bf67-kcmj8"] Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.883139 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-config\") pod \"453feb9e-4bbd-49b8-9719-07f97af54b46\" (UID: \"453feb9e-4bbd-49b8-9719-07f97af54b46\") " Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.883750 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-config" (OuterVolumeSpecName: "config") pod "453feb9e-4bbd-49b8-9719-07f97af54b46" (UID: "453feb9e-4bbd-49b8-9719-07f97af54b46"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:49:02 crc kubenswrapper[4894]: I1209 15:49:02.883894 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/453feb9e-4bbd-49b8-9719-07f97af54b46-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:03 crc kubenswrapper[4894]: I1209 15:49:03.225053 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-649c9bf67-kcmj8" event={"ID":"18cc81c1-7bef-41c2-89b7-e507624504f4","Type":"ContainerStarted","Data":"eeaedfae59c5b4855abafb14f13be9032c198decab6a35bd3a3cd39ef705e2d4"} Dec 09 15:49:03 crc kubenswrapper[4894]: I1209 15:49:03.227984 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-66dd8c6975-c6tff" event={"ID":"453feb9e-4bbd-49b8-9719-07f97af54b46","Type":"ContainerDied","Data":"357d71490f25db3c3df0259eefc713c4d4afc1148392030d8825cc8b1985aa8e"} Dec 09 15:49:03 crc kubenswrapper[4894]: I1209 15:49:03.228042 4894 scope.go:117] "RemoveContainer" containerID="166be08764e066427ee921580f9db9c6eeb23b7a34359ff64e1dd0a93ae54ae2" Dec 09 15:49:03 crc kubenswrapper[4894]: I1209 15:49:03.228151 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-66dd8c6975-c6tff" Dec 09 15:49:03 crc kubenswrapper[4894]: I1209 15:49:03.247319 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" event={"ID":"f20c656d-4752-41ce-a1aa-aa6980430e90","Type":"ContainerStarted","Data":"09c02c60b86e32b44013a0492a40353ce279985637cd09e8deeb9bbcfac8cbbb"} Dec 09 15:49:03 crc kubenswrapper[4894]: I1209 15:49:03.247385 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" Dec 09 15:49:03 crc kubenswrapper[4894]: I1209 15:49:03.267797 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" podStartSLOduration=4.267781132 podStartE2EDuration="4.267781132s" podCreationTimestamp="2025-12-09 15:48:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:49:03.26654489 +0000 UTC m=+1037.585755559" watchObservedRunningTime="2025-12-09 15:49:03.267781132 +0000 UTC m=+1037.586991801" Dec 09 15:49:03 crc kubenswrapper[4894]: I1209 15:49:03.353870 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-66dd8c6975-c6tff"] Dec 09 15:49:03 crc kubenswrapper[4894]: I1209 15:49:03.361596 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-66dd8c6975-c6tff"] Dec 09 15:49:04 crc kubenswrapper[4894]: I1209 15:49:04.117744 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="453feb9e-4bbd-49b8-9719-07f97af54b46" path="/var/lib/kubelet/pods/453feb9e-4bbd-49b8-9719-07f97af54b46/volumes" Dec 09 15:49:05 crc kubenswrapper[4894]: I1209 15:49:05.267811 4894 generic.go:334] "Generic (PLEG): container finished" podID="ae433ea4-7305-4d95-96b5-da695d0ed504" containerID="466350b715f0a001b96f8b05271b4b5a567f93e5f14fc6dd76ac361dc90e20d6" exitCode=0 Dec 09 15:49:05 crc kubenswrapper[4894]: I1209 15:49:05.267862 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-274hf" event={"ID":"ae433ea4-7305-4d95-96b5-da695d0ed504","Type":"ContainerDied","Data":"466350b715f0a001b96f8b05271b4b5a567f93e5f14fc6dd76ac361dc90e20d6"} Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.196533 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6ff8f88b5f-gnrzd"] Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.219196 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-56f65bfbb-tv5ld"] Dec 09 15:49:08 crc kubenswrapper[4894]: E1209 15:49:08.220443 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="453feb9e-4bbd-49b8-9719-07f97af54b46" containerName="init" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.220459 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="453feb9e-4bbd-49b8-9719-07f97af54b46" containerName="init" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.221113 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="453feb9e-4bbd-49b8-9719-07f97af54b46" containerName="init" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.222028 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.227250 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-horizon-svc" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.247296 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-56f65bfbb-tv5ld"] Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.294061 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-scripts\") pod \"horizon-56f65bfbb-tv5ld\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.294111 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-horizon-secret-key\") pod \"horizon-56f65bfbb-tv5ld\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.294168 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-config-data\") pod \"horizon-56f65bfbb-tv5ld\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.294248 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-combined-ca-bundle\") pod \"horizon-56f65bfbb-tv5ld\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.294275 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-horizon-tls-certs\") pod \"horizon-56f65bfbb-tv5ld\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.294293 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d65gh\" (UniqueName: \"kubernetes.io/projected/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-kube-api-access-d65gh\") pod \"horizon-56f65bfbb-tv5ld\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.294311 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-logs\") pod \"horizon-56f65bfbb-tv5ld\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.347722 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-649c9bf67-kcmj8"] Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.370001 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizon-6b9bcbff6b-l6n7z"] Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.372153 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.396602 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-scripts\") pod \"horizon-56f65bfbb-tv5ld\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.396662 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-horizon-secret-key\") pod \"horizon-56f65bfbb-tv5ld\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.396701 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-config-data\") pod \"horizon-56f65bfbb-tv5ld\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.396763 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-combined-ca-bundle\") pod \"horizon-56f65bfbb-tv5ld\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.396787 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-horizon-tls-certs\") pod \"horizon-56f65bfbb-tv5ld\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.396807 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d65gh\" (UniqueName: \"kubernetes.io/projected/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-kube-api-access-d65gh\") pod \"horizon-56f65bfbb-tv5ld\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.396826 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-logs\") pod \"horizon-56f65bfbb-tv5ld\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.397282 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-logs\") pod \"horizon-56f65bfbb-tv5ld\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.398244 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-config-data\") pod \"horizon-56f65bfbb-tv5ld\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.398610 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-scripts\") pod \"horizon-56f65bfbb-tv5ld\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.400111 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6b9bcbff6b-l6n7z"] Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.405264 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-combined-ca-bundle\") pod \"horizon-56f65bfbb-tv5ld\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.405330 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-horizon-tls-certs\") pod \"horizon-56f65bfbb-tv5ld\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.406178 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-horizon-secret-key\") pod \"horizon-56f65bfbb-tv5ld\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.426062 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d65gh\" (UniqueName: \"kubernetes.io/projected/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-kube-api-access-d65gh\") pod \"horizon-56f65bfbb-tv5ld\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.498936 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6b8de55a-230d-4dbe-9f6c-aae6138d865f-horizon-secret-key\") pod \"horizon-6b9bcbff6b-l6n7z\" (UID: \"6b8de55a-230d-4dbe-9f6c-aae6138d865f\") " pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.498993 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b8de55a-230d-4dbe-9f6c-aae6138d865f-horizon-tls-certs\") pod \"horizon-6b9bcbff6b-l6n7z\" (UID: \"6b8de55a-230d-4dbe-9f6c-aae6138d865f\") " pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.499027 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qdpvq\" (UniqueName: \"kubernetes.io/projected/6b8de55a-230d-4dbe-9f6c-aae6138d865f-kube-api-access-qdpvq\") pod \"horizon-6b9bcbff6b-l6n7z\" (UID: \"6b8de55a-230d-4dbe-9f6c-aae6138d865f\") " pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.499095 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6b8de55a-230d-4dbe-9f6c-aae6138d865f-config-data\") pod \"horizon-6b9bcbff6b-l6n7z\" (UID: \"6b8de55a-230d-4dbe-9f6c-aae6138d865f\") " pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.499130 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b8de55a-230d-4dbe-9f6c-aae6138d865f-scripts\") pod \"horizon-6b9bcbff6b-l6n7z\" (UID: \"6b8de55a-230d-4dbe-9f6c-aae6138d865f\") " pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.499151 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8de55a-230d-4dbe-9f6c-aae6138d865f-combined-ca-bundle\") pod \"horizon-6b9bcbff6b-l6n7z\" (UID: \"6b8de55a-230d-4dbe-9f6c-aae6138d865f\") " pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.499181 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b8de55a-230d-4dbe-9f6c-aae6138d865f-logs\") pod \"horizon-6b9bcbff6b-l6n7z\" (UID: \"6b8de55a-230d-4dbe-9f6c-aae6138d865f\") " pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.563057 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.601400 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b8de55a-230d-4dbe-9f6c-aae6138d865f-scripts\") pod \"horizon-6b9bcbff6b-l6n7z\" (UID: \"6b8de55a-230d-4dbe-9f6c-aae6138d865f\") " pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.601450 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8de55a-230d-4dbe-9f6c-aae6138d865f-combined-ca-bundle\") pod \"horizon-6b9bcbff6b-l6n7z\" (UID: \"6b8de55a-230d-4dbe-9f6c-aae6138d865f\") " pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.601487 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b8de55a-230d-4dbe-9f6c-aae6138d865f-logs\") pod \"horizon-6b9bcbff6b-l6n7z\" (UID: \"6b8de55a-230d-4dbe-9f6c-aae6138d865f\") " pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.601539 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6b8de55a-230d-4dbe-9f6c-aae6138d865f-horizon-secret-key\") pod \"horizon-6b9bcbff6b-l6n7z\" (UID: \"6b8de55a-230d-4dbe-9f6c-aae6138d865f\") " pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.602394 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6b8de55a-230d-4dbe-9f6c-aae6138d865f-logs\") pod \"horizon-6b9bcbff6b-l6n7z\" (UID: \"6b8de55a-230d-4dbe-9f6c-aae6138d865f\") " pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.601571 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b8de55a-230d-4dbe-9f6c-aae6138d865f-horizon-tls-certs\") pod \"horizon-6b9bcbff6b-l6n7z\" (UID: \"6b8de55a-230d-4dbe-9f6c-aae6138d865f\") " pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.602682 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/6b8de55a-230d-4dbe-9f6c-aae6138d865f-scripts\") pod \"horizon-6b9bcbff6b-l6n7z\" (UID: \"6b8de55a-230d-4dbe-9f6c-aae6138d865f\") " pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.602762 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qdpvq\" (UniqueName: \"kubernetes.io/projected/6b8de55a-230d-4dbe-9f6c-aae6138d865f-kube-api-access-qdpvq\") pod \"horizon-6b9bcbff6b-l6n7z\" (UID: \"6b8de55a-230d-4dbe-9f6c-aae6138d865f\") " pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.602855 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6b8de55a-230d-4dbe-9f6c-aae6138d865f-config-data\") pod \"horizon-6b9bcbff6b-l6n7z\" (UID: \"6b8de55a-230d-4dbe-9f6c-aae6138d865f\") " pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.604042 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/6b8de55a-230d-4dbe-9f6c-aae6138d865f-config-data\") pod \"horizon-6b9bcbff6b-l6n7z\" (UID: \"6b8de55a-230d-4dbe-9f6c-aae6138d865f\") " pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.610826 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6b8de55a-230d-4dbe-9f6c-aae6138d865f-combined-ca-bundle\") pod \"horizon-6b9bcbff6b-l6n7z\" (UID: \"6b8de55a-230d-4dbe-9f6c-aae6138d865f\") " pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.613337 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/6b8de55a-230d-4dbe-9f6c-aae6138d865f-horizon-tls-certs\") pod \"horizon-6b9bcbff6b-l6n7z\" (UID: \"6b8de55a-230d-4dbe-9f6c-aae6138d865f\") " pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.613686 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/6b8de55a-230d-4dbe-9f6c-aae6138d865f-horizon-secret-key\") pod \"horizon-6b9bcbff6b-l6n7z\" (UID: \"6b8de55a-230d-4dbe-9f6c-aae6138d865f\") " pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.623245 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qdpvq\" (UniqueName: \"kubernetes.io/projected/6b8de55a-230d-4dbe-9f6c-aae6138d865f-kube-api-access-qdpvq\") pod \"horizon-6b9bcbff6b-l6n7z\" (UID: \"6b8de55a-230d-4dbe-9f6c-aae6138d865f\") " pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:08 crc kubenswrapper[4894]: I1209 15:49:08.690858 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:09 crc kubenswrapper[4894]: I1209 15:49:09.839702 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-274hf" Dec 09 15:49:09 crc kubenswrapper[4894]: I1209 15:49:09.930437 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-scripts\") pod \"ae433ea4-7305-4d95-96b5-da695d0ed504\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " Dec 09 15:49:09 crc kubenswrapper[4894]: I1209 15:49:09.930511 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-combined-ca-bundle\") pod \"ae433ea4-7305-4d95-96b5-da695d0ed504\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " Dec 09 15:49:09 crc kubenswrapper[4894]: I1209 15:49:09.930658 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-fernet-keys\") pod \"ae433ea4-7305-4d95-96b5-da695d0ed504\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " Dec 09 15:49:09 crc kubenswrapper[4894]: I1209 15:49:09.930704 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-config-data\") pod \"ae433ea4-7305-4d95-96b5-da695d0ed504\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " Dec 09 15:49:09 crc kubenswrapper[4894]: I1209 15:49:09.930795 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-credential-keys\") pod \"ae433ea4-7305-4d95-96b5-da695d0ed504\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " Dec 09 15:49:09 crc kubenswrapper[4894]: I1209 15:49:09.930873 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t5ldj\" (UniqueName: \"kubernetes.io/projected/ae433ea4-7305-4d95-96b5-da695d0ed504-kube-api-access-t5ldj\") pod \"ae433ea4-7305-4d95-96b5-da695d0ed504\" (UID: \"ae433ea4-7305-4d95-96b5-da695d0ed504\") " Dec 09 15:49:09 crc kubenswrapper[4894]: I1209 15:49:09.936290 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "ae433ea4-7305-4d95-96b5-da695d0ed504" (UID: "ae433ea4-7305-4d95-96b5-da695d0ed504"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:09 crc kubenswrapper[4894]: I1209 15:49:09.936327 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ae433ea4-7305-4d95-96b5-da695d0ed504-kube-api-access-t5ldj" (OuterVolumeSpecName: "kube-api-access-t5ldj") pod "ae433ea4-7305-4d95-96b5-da695d0ed504" (UID: "ae433ea4-7305-4d95-96b5-da695d0ed504"). InnerVolumeSpecName "kube-api-access-t5ldj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:49:09 crc kubenswrapper[4894]: I1209 15:49:09.936691 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-scripts" (OuterVolumeSpecName: "scripts") pod "ae433ea4-7305-4d95-96b5-da695d0ed504" (UID: "ae433ea4-7305-4d95-96b5-da695d0ed504"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:09 crc kubenswrapper[4894]: I1209 15:49:09.938238 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "ae433ea4-7305-4d95-96b5-da695d0ed504" (UID: "ae433ea4-7305-4d95-96b5-da695d0ed504"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:09 crc kubenswrapper[4894]: I1209 15:49:09.971729 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ae433ea4-7305-4d95-96b5-da695d0ed504" (UID: "ae433ea4-7305-4d95-96b5-da695d0ed504"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:09 crc kubenswrapper[4894]: I1209 15:49:09.984233 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-config-data" (OuterVolumeSpecName: "config-data") pod "ae433ea4-7305-4d95-96b5-da695d0ed504" (UID: "ae433ea4-7305-4d95-96b5-da695d0ed504"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:10 crc kubenswrapper[4894]: I1209 15:49:10.032705 4894 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:10 crc kubenswrapper[4894]: I1209 15:49:10.032737 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:10 crc kubenswrapper[4894]: I1209 15:49:10.032746 4894 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:10 crc kubenswrapper[4894]: I1209 15:49:10.032758 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t5ldj\" (UniqueName: \"kubernetes.io/projected/ae433ea4-7305-4d95-96b5-da695d0ed504-kube-api-access-t5ldj\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:10 crc kubenswrapper[4894]: I1209 15:49:10.032766 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:10 crc kubenswrapper[4894]: I1209 15:49:10.032775 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ae433ea4-7305-4d95-96b5-da695d0ed504-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:10 crc kubenswrapper[4894]: I1209 15:49:10.323668 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" Dec 09 15:49:10 crc kubenswrapper[4894]: I1209 15:49:10.339344 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-274hf" event={"ID":"ae433ea4-7305-4d95-96b5-da695d0ed504","Type":"ContainerDied","Data":"8bbf03beb82f7a65c3c169b766814f6fd1c6d072354bd302596626387be7bb82"} Dec 09 15:49:10 crc kubenswrapper[4894]: I1209 15:49:10.339418 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8bbf03beb82f7a65c3c169b766814f6fd1c6d072354bd302596626387be7bb82" Dec 09 15:49:10 crc kubenswrapper[4894]: I1209 15:49:10.339551 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-274hf" Dec 09 15:49:10 crc kubenswrapper[4894]: I1209 15:49:10.479826 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bc56d6f79-xg28g"] Dec 09 15:49:10 crc kubenswrapper[4894]: I1209 15:49:10.480740 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" podUID="5c6eec80-a266-4965-9b10-c5e7b9dee04d" containerName="dnsmasq-dns" containerID="cri-o://bd664398729728de264c1e429a821521b44bbc925e4855a7292f30c11396b30e" gracePeriod=10 Dec 09 15:49:10 crc kubenswrapper[4894]: I1209 15:49:10.924795 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-274hf"] Dec 09 15:49:10 crc kubenswrapper[4894]: I1209 15:49:10.932263 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-274hf"] Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.025030 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-rl92n"] Dec 09 15:49:11 crc kubenswrapper[4894]: E1209 15:49:11.025489 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ae433ea4-7305-4d95-96b5-da695d0ed504" containerName="keystone-bootstrap" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.025521 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ae433ea4-7305-4d95-96b5-da695d0ed504" containerName="keystone-bootstrap" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.025815 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="ae433ea4-7305-4d95-96b5-da695d0ed504" containerName="keystone-bootstrap" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.026742 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rl92n" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.031056 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.031494 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-g8fmh" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.031866 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.032012 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.033843 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.041922 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rl92n"] Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.173706 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-combined-ca-bundle\") pod \"keystone-bootstrap-rl92n\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " pod="openstack/keystone-bootstrap-rl92n" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.174099 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-scripts\") pod \"keystone-bootstrap-rl92n\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " pod="openstack/keystone-bootstrap-rl92n" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.174152 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-credential-keys\") pod \"keystone-bootstrap-rl92n\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " pod="openstack/keystone-bootstrap-rl92n" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.174191 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-config-data\") pod \"keystone-bootstrap-rl92n\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " pod="openstack/keystone-bootstrap-rl92n" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.174315 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mlb5l\" (UniqueName: \"kubernetes.io/projected/febdd308-ca10-4ba3-a34c-b000517eaeeb-kube-api-access-mlb5l\") pod \"keystone-bootstrap-rl92n\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " pod="openstack/keystone-bootstrap-rl92n" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.174346 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-fernet-keys\") pod \"keystone-bootstrap-rl92n\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " pod="openstack/keystone-bootstrap-rl92n" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.275730 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mlb5l\" (UniqueName: \"kubernetes.io/projected/febdd308-ca10-4ba3-a34c-b000517eaeeb-kube-api-access-mlb5l\") pod \"keystone-bootstrap-rl92n\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " pod="openstack/keystone-bootstrap-rl92n" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.275779 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-fernet-keys\") pod \"keystone-bootstrap-rl92n\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " pod="openstack/keystone-bootstrap-rl92n" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.275832 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-combined-ca-bundle\") pod \"keystone-bootstrap-rl92n\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " pod="openstack/keystone-bootstrap-rl92n" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.275860 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-scripts\") pod \"keystone-bootstrap-rl92n\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " pod="openstack/keystone-bootstrap-rl92n" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.275884 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-credential-keys\") pod \"keystone-bootstrap-rl92n\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " pod="openstack/keystone-bootstrap-rl92n" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.275909 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-config-data\") pod \"keystone-bootstrap-rl92n\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " pod="openstack/keystone-bootstrap-rl92n" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.282160 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-combined-ca-bundle\") pod \"keystone-bootstrap-rl92n\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " pod="openstack/keystone-bootstrap-rl92n" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.282268 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-fernet-keys\") pod \"keystone-bootstrap-rl92n\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " pod="openstack/keystone-bootstrap-rl92n" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.283272 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-scripts\") pod \"keystone-bootstrap-rl92n\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " pod="openstack/keystone-bootstrap-rl92n" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.283276 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-credential-keys\") pod \"keystone-bootstrap-rl92n\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " pod="openstack/keystone-bootstrap-rl92n" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.292713 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-config-data\") pod \"keystone-bootstrap-rl92n\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " pod="openstack/keystone-bootstrap-rl92n" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.293086 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mlb5l\" (UniqueName: \"kubernetes.io/projected/febdd308-ca10-4ba3-a34c-b000517eaeeb-kube-api-access-mlb5l\") pod \"keystone-bootstrap-rl92n\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " pod="openstack/keystone-bootstrap-rl92n" Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.353389 4894 generic.go:334] "Generic (PLEG): container finished" podID="5c6eec80-a266-4965-9b10-c5e7b9dee04d" containerID="bd664398729728de264c1e429a821521b44bbc925e4855a7292f30c11396b30e" exitCode=0 Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.353439 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" event={"ID":"5c6eec80-a266-4965-9b10-c5e7b9dee04d","Type":"ContainerDied","Data":"bd664398729728de264c1e429a821521b44bbc925e4855a7292f30c11396b30e"} Dec 09 15:49:11 crc kubenswrapper[4894]: I1209 15:49:11.363036 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rl92n" Dec 09 15:49:12 crc kubenswrapper[4894]: I1209 15:49:12.117278 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ae433ea4-7305-4d95-96b5-da695d0ed504" path="/var/lib/kubelet/pods/ae433ea4-7305-4d95-96b5-da695d0ed504/volumes" Dec 09 15:49:12 crc kubenswrapper[4894]: I1209 15:49:12.274283 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" podUID="5c6eec80-a266-4965-9b10-c5e7b9dee04d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.122:5353: connect: connection refused" Dec 09 15:49:17 crc kubenswrapper[4894]: I1209 15:49:17.274622 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" podUID="5c6eec80-a266-4965-9b10-c5e7b9dee04d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.122:5353: connect: connection refused" Dec 09 15:49:20 crc kubenswrapper[4894]: E1209 15:49:20.742997 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:dd7600bc5278c663cfcfecafd3fb051a2cd2ddc3c1efb07738bf09512aa23ae7" Dec 09 15:49:20 crc kubenswrapper[4894]: E1209 15:49:20.743786 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:dd7600bc5278c663cfcfecafd3fb051a2cd2ddc3c1efb07738bf09512aa23ae7,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n5bdh567h555h76h8fh68fh686h659hc4h95h5b7h58fhb9h58h78h96h66dh58bh5d4hb8h668h558h6dh649h5b8h5d7h68ch65h698h5dbh54fh6dq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-bvtfd,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-6ff8f88b5f-gnrzd_openstack(1ad9eb8c-d035-456d-a345-5a4da859fc7c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 15:49:20 crc kubenswrapper[4894]: E1209 15:49:20.748232 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:dd7600bc5278c663cfcfecafd3fb051a2cd2ddc3c1efb07738bf09512aa23ae7\\\"\"]" pod="openstack/horizon-6ff8f88b5f-gnrzd" podUID="1ad9eb8c-d035-456d-a345-5a4da859fc7c" Dec 09 15:49:22 crc kubenswrapper[4894]: I1209 15:49:22.274443 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" podUID="5c6eec80-a266-4965-9b10-c5e7b9dee04d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.122:5353: connect: connection refused" Dec 09 15:49:22 crc kubenswrapper[4894]: I1209 15:49:22.274896 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" Dec 09 15:49:22 crc kubenswrapper[4894]: E1209 15:49:22.427280 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:dd7600bc5278c663cfcfecafd3fb051a2cd2ddc3c1efb07738bf09512aa23ae7" Dec 09 15:49:22 crc kubenswrapper[4894]: E1209 15:49:22.427463 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:dd7600bc5278c663cfcfecafd3fb051a2cd2ddc3c1efb07738bf09512aa23ae7,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n698h5cbh655h99hf8h8h5b4h5cdh564h676h657h546h547h646h549h5dch6fh64ch54fh79h5b9h59ch87h677h66ch67dh9ch5dbh697h74hdch57bq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pq92x,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-649c9bf67-kcmj8_openstack(18cc81c1-7bef-41c2-89b7-e507624504f4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 15:49:22 crc kubenswrapper[4894]: E1209 15:49:22.430841 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizon@sha256:dd7600bc5278c663cfcfecafd3fb051a2cd2ddc3c1efb07738bf09512aa23ae7" Dec 09 15:49:22 crc kubenswrapper[4894]: E1209 15:49:22.431074 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizon-log,Image:quay.io/podified-antelope-centos9/openstack-horizon@sha256:dd7600bc5278c663cfcfecafd3fb051a2cd2ddc3c1efb07738bf09512aa23ae7,Command:[/bin/bash],Args:[-c tail -n+1 -F /var/log/horizon/horizon.log],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:n595h687hcch5d7hbh7bh99hb8h58bh596h569h68fh6ch6h54fh548hd6hfdh669hf8h58dh5h674h667h68ch589hb5h75h57ch87h678h65cq,ValueFrom:nil,},EnvVar{Name:ENABLE_DESIGNATE,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_HEAT,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_IRONIC,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_MANILA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_OCTAVIA,Value:yes,ValueFrom:nil,},EnvVar{Name:ENABLE_WATCHER,Value:no,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:UNPACK_THEME,Value:true,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:logs,ReadOnly:false,MountPath:/var/log/horizon,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-xm8mg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*48,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*true,RunAsGroup:*42400,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizon-55fff446b9-mbvp9_openstack(ff441522-39b1-47e3-be7d-d590baec8889): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 15:49:22 crc kubenswrapper[4894]: E1209 15:49:22.439748 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:dd7600bc5278c663cfcfecafd3fb051a2cd2ddc3c1efb07738bf09512aa23ae7\\\"\"]" pod="openstack/horizon-649c9bf67-kcmj8" podUID="18cc81c1-7bef-41c2-89b7-e507624504f4" Dec 09 15:49:22 crc kubenswrapper[4894]: E1209 15:49:22.440033 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="[failed to \"StartContainer\" for \"horizon-log\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\", failed to \"StartContainer\" for \"horizon\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizon@sha256:dd7600bc5278c663cfcfecafd3fb051a2cd2ddc3c1efb07738bf09512aa23ae7\\\"\"]" pod="openstack/horizon-55fff446b9-mbvp9" podUID="ff441522-39b1-47e3-be7d-d590baec8889" Dec 09 15:49:28 crc kubenswrapper[4894]: I1209 15:49:28.544946 4894 generic.go:334] "Generic (PLEG): container finished" podID="b48430a2-ef1d-4747-97a4-4b6933fcea4f" containerID="b8a1f501c7d2757a2ff9a88f8716626e02ee6c30fae871162ab553918ea821dc" exitCode=0 Dec 09 15:49:28 crc kubenswrapper[4894]: I1209 15:49:28.545042 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lfx27" event={"ID":"b48430a2-ef1d-4747-97a4-4b6933fcea4f","Type":"ContainerDied","Data":"b8a1f501c7d2757a2ff9a88f8716626e02ee6c30fae871162ab553918ea821dc"} Dec 09 15:49:32 crc kubenswrapper[4894]: E1209 15:49:32.033917 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:fe32d3ea620f0c7ecfdde9bbf28417fde03bc18c6f60b1408fa8da24d8188f16" Dec 09 15:49:32 crc kubenswrapper[4894]: E1209 15:49:32.034405 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:barbican-db-sync,Image:quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:fe32d3ea620f0c7ecfdde9bbf28417fde03bc18c6f60b1408fa8da24d8188f16,Command:[/bin/bash],Args:[-c barbican-manage db upgrade],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/barbican/barbican.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-b4rm6,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42403,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42403,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-db-sync-s4wr5_openstack(d29fb637-ba21-4cf9-ae4d-cd9720dd5af4): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 15:49:32 crc kubenswrapper[4894]: E1209 15:49:32.035592 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/barbican-db-sync-s4wr5" podUID="d29fb637-ba21-4cf9-ae4d-cd9720dd5af4" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.179197 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6ff8f88b5f-gnrzd" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.198444 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-649c9bf67-kcmj8" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.203253 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.205695 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lfx27" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.208666 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55fff446b9-mbvp9" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.215575 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvtfd\" (UniqueName: \"kubernetes.io/projected/1ad9eb8c-d035-456d-a345-5a4da859fc7c-kube-api-access-bvtfd\") pod \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\" (UID: \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.215610 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ad9eb8c-d035-456d-a345-5a4da859fc7c-config-data\") pod \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\" (UID: \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.215768 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ad9eb8c-d035-456d-a345-5a4da859fc7c-logs\") pod \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\" (UID: \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.215844 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ad9eb8c-d035-456d-a345-5a4da859fc7c-scripts\") pod \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\" (UID: \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.215895 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1ad9eb8c-d035-456d-a345-5a4da859fc7c-horizon-secret-key\") pod \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\" (UID: \"1ad9eb8c-d035-456d-a345-5a4da859fc7c\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.221207 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ad9eb8c-d035-456d-a345-5a4da859fc7c-config-data" (OuterVolumeSpecName: "config-data") pod "1ad9eb8c-d035-456d-a345-5a4da859fc7c" (UID: "1ad9eb8c-d035-456d-a345-5a4da859fc7c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.221819 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ad9eb8c-d035-456d-a345-5a4da859fc7c-logs" (OuterVolumeSpecName: "logs") pod "1ad9eb8c-d035-456d-a345-5a4da859fc7c" (UID: "1ad9eb8c-d035-456d-a345-5a4da859fc7c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.222468 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1ad9eb8c-d035-456d-a345-5a4da859fc7c-scripts" (OuterVolumeSpecName: "scripts") pod "1ad9eb8c-d035-456d-a345-5a4da859fc7c" (UID: "1ad9eb8c-d035-456d-a345-5a4da859fc7c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.225369 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ad9eb8c-d035-456d-a345-5a4da859fc7c-kube-api-access-bvtfd" (OuterVolumeSpecName: "kube-api-access-bvtfd") pod "1ad9eb8c-d035-456d-a345-5a4da859fc7c" (UID: "1ad9eb8c-d035-456d-a345-5a4da859fc7c"). InnerVolumeSpecName "kube-api-access-bvtfd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.229807 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ad9eb8c-d035-456d-a345-5a4da859fc7c-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "1ad9eb8c-d035-456d-a345-5a4da859fc7c" (UID: "1ad9eb8c-d035-456d-a345-5a4da859fc7c"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.275007 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" podUID="5c6eec80-a266-4965-9b10-c5e7b9dee04d" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.122:5353: i/o timeout" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.317604 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-ovsdbserver-nb\") pod \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\" (UID: \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.317662 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggwmj\" (UniqueName: \"kubernetes.io/projected/b48430a2-ef1d-4747-97a4-4b6933fcea4f-kube-api-access-ggwmj\") pod \"b48430a2-ef1d-4747-97a4-4b6933fcea4f\" (UID: \"b48430a2-ef1d-4747-97a4-4b6933fcea4f\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.317686 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/18cc81c1-7bef-41c2-89b7-e507624504f4-horizon-secret-key\") pod \"18cc81c1-7bef-41c2-89b7-e507624504f4\" (UID: \"18cc81c1-7bef-41c2-89b7-e507624504f4\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.317730 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff441522-39b1-47e3-be7d-d590baec8889-logs\") pod \"ff441522-39b1-47e3-be7d-d590baec8889\" (UID: \"ff441522-39b1-47e3-be7d-d590baec8889\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.317752 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18cc81c1-7bef-41c2-89b7-e507624504f4-logs\") pod \"18cc81c1-7bef-41c2-89b7-e507624504f4\" (UID: \"18cc81c1-7bef-41c2-89b7-e507624504f4\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.317783 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18cc81c1-7bef-41c2-89b7-e507624504f4-scripts\") pod \"18cc81c1-7bef-41c2-89b7-e507624504f4\" (UID: \"18cc81c1-7bef-41c2-89b7-e507624504f4\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.317840 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-config\") pod \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\" (UID: \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.317866 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/b48430a2-ef1d-4747-97a4-4b6933fcea4f-config\") pod \"b48430a2-ef1d-4747-97a4-4b6933fcea4f\" (UID: \"b48430a2-ef1d-4747-97a4-4b6933fcea4f\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.317905 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b48430a2-ef1d-4747-97a4-4b6933fcea4f-combined-ca-bundle\") pod \"b48430a2-ef1d-4747-97a4-4b6933fcea4f\" (UID: \"b48430a2-ef1d-4747-97a4-4b6933fcea4f\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.317925 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-ovsdbserver-sb\") pod \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\" (UID: \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.317943 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xm8mg\" (UniqueName: \"kubernetes.io/projected/ff441522-39b1-47e3-be7d-d590baec8889-kube-api-access-xm8mg\") pod \"ff441522-39b1-47e3-be7d-d590baec8889\" (UID: \"ff441522-39b1-47e3-be7d-d590baec8889\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.317966 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff441522-39b1-47e3-be7d-d590baec8889-config-data\") pod \"ff441522-39b1-47e3-be7d-d590baec8889\" (UID: \"ff441522-39b1-47e3-be7d-d590baec8889\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.318008 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p4m8h\" (UniqueName: \"kubernetes.io/projected/5c6eec80-a266-4965-9b10-c5e7b9dee04d-kube-api-access-p4m8h\") pod \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\" (UID: \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.318024 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18cc81c1-7bef-41c2-89b7-e507624504f4-config-data\") pod \"18cc81c1-7bef-41c2-89b7-e507624504f4\" (UID: \"18cc81c1-7bef-41c2-89b7-e507624504f4\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.318045 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff441522-39b1-47e3-be7d-d590baec8889-scripts\") pod \"ff441522-39b1-47e3-be7d-d590baec8889\" (UID: \"ff441522-39b1-47e3-be7d-d590baec8889\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.318084 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-dns-svc\") pod \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\" (UID: \"5c6eec80-a266-4965-9b10-c5e7b9dee04d\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.318115 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pq92x\" (UniqueName: \"kubernetes.io/projected/18cc81c1-7bef-41c2-89b7-e507624504f4-kube-api-access-pq92x\") pod \"18cc81c1-7bef-41c2-89b7-e507624504f4\" (UID: \"18cc81c1-7bef-41c2-89b7-e507624504f4\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.318147 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ff441522-39b1-47e3-be7d-d590baec8889-horizon-secret-key\") pod \"ff441522-39b1-47e3-be7d-d590baec8889\" (UID: \"ff441522-39b1-47e3-be7d-d590baec8889\") " Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.318484 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/1ad9eb8c-d035-456d-a345-5a4da859fc7c-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.318497 4894 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/1ad9eb8c-d035-456d-a345-5a4da859fc7c-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.318510 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvtfd\" (UniqueName: \"kubernetes.io/projected/1ad9eb8c-d035-456d-a345-5a4da859fc7c-kube-api-access-bvtfd\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.318518 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/1ad9eb8c-d035-456d-a345-5a4da859fc7c-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.318527 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ad9eb8c-d035-456d-a345-5a4da859fc7c-logs\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.320307 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ff441522-39b1-47e3-be7d-d590baec8889-logs" (OuterVolumeSpecName: "logs") pod "ff441522-39b1-47e3-be7d-d590baec8889" (UID: "ff441522-39b1-47e3-be7d-d590baec8889"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.322669 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18cc81c1-7bef-41c2-89b7-e507624504f4-scripts" (OuterVolumeSpecName: "scripts") pod "18cc81c1-7bef-41c2-89b7-e507624504f4" (UID: "18cc81c1-7bef-41c2-89b7-e507624504f4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.323093 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/18cc81c1-7bef-41c2-89b7-e507624504f4-logs" (OuterVolumeSpecName: "logs") pod "18cc81c1-7bef-41c2-89b7-e507624504f4" (UID: "18cc81c1-7bef-41c2-89b7-e507624504f4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.323281 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff441522-39b1-47e3-be7d-d590baec8889-scripts" (OuterVolumeSpecName: "scripts") pod "ff441522-39b1-47e3-be7d-d590baec8889" (UID: "ff441522-39b1-47e3-be7d-d590baec8889"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.326712 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/18cc81c1-7bef-41c2-89b7-e507624504f4-config-data" (OuterVolumeSpecName: "config-data") pod "18cc81c1-7bef-41c2-89b7-e507624504f4" (UID: "18cc81c1-7bef-41c2-89b7-e507624504f4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.328830 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ff441522-39b1-47e3-be7d-d590baec8889-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "ff441522-39b1-47e3-be7d-d590baec8889" (UID: "ff441522-39b1-47e3-be7d-d590baec8889"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.329871 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ff441522-39b1-47e3-be7d-d590baec8889-config-data" (OuterVolumeSpecName: "config-data") pod "ff441522-39b1-47e3-be7d-d590baec8889" (UID: "ff441522-39b1-47e3-be7d-d590baec8889"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.332263 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/18cc81c1-7bef-41c2-89b7-e507624504f4-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "18cc81c1-7bef-41c2-89b7-e507624504f4" (UID: "18cc81c1-7bef-41c2-89b7-e507624504f4"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.333264 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c6eec80-a266-4965-9b10-c5e7b9dee04d-kube-api-access-p4m8h" (OuterVolumeSpecName: "kube-api-access-p4m8h") pod "5c6eec80-a266-4965-9b10-c5e7b9dee04d" (UID: "5c6eec80-a266-4965-9b10-c5e7b9dee04d"). InnerVolumeSpecName "kube-api-access-p4m8h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.346959 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b48430a2-ef1d-4747-97a4-4b6933fcea4f-kube-api-access-ggwmj" (OuterVolumeSpecName: "kube-api-access-ggwmj") pod "b48430a2-ef1d-4747-97a4-4b6933fcea4f" (UID: "b48430a2-ef1d-4747-97a4-4b6933fcea4f"). InnerVolumeSpecName "kube-api-access-ggwmj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.347073 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ff441522-39b1-47e3-be7d-d590baec8889-kube-api-access-xm8mg" (OuterVolumeSpecName: "kube-api-access-xm8mg") pod "ff441522-39b1-47e3-be7d-d590baec8889" (UID: "ff441522-39b1-47e3-be7d-d590baec8889"). InnerVolumeSpecName "kube-api-access-xm8mg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.349528 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/18cc81c1-7bef-41c2-89b7-e507624504f4-kube-api-access-pq92x" (OuterVolumeSpecName: "kube-api-access-pq92x") pod "18cc81c1-7bef-41c2-89b7-e507624504f4" (UID: "18cc81c1-7bef-41c2-89b7-e507624504f4"). InnerVolumeSpecName "kube-api-access-pq92x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.354669 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b48430a2-ef1d-4747-97a4-4b6933fcea4f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b48430a2-ef1d-4747-97a4-4b6933fcea4f" (UID: "b48430a2-ef1d-4747-97a4-4b6933fcea4f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.371627 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b48430a2-ef1d-4747-97a4-4b6933fcea4f-config" (OuterVolumeSpecName: "config") pod "b48430a2-ef1d-4747-97a4-4b6933fcea4f" (UID: "b48430a2-ef1d-4747-97a4-4b6933fcea4f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.374605 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "5c6eec80-a266-4965-9b10-c5e7b9dee04d" (UID: "5c6eec80-a266-4965-9b10-c5e7b9dee04d"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.378888 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-config" (OuterVolumeSpecName: "config") pod "5c6eec80-a266-4965-9b10-c5e7b9dee04d" (UID: "5c6eec80-a266-4965-9b10-c5e7b9dee04d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.383523 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "5c6eec80-a266-4965-9b10-c5e7b9dee04d" (UID: "5c6eec80-a266-4965-9b10-c5e7b9dee04d"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.391003 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "5c6eec80-a266-4965-9b10-c5e7b9dee04d" (UID: "5c6eec80-a266-4965-9b10-c5e7b9dee04d"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.420620 4894 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/ff441522-39b1-47e3-be7d-d590baec8889-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.420679 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.420690 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggwmj\" (UniqueName: \"kubernetes.io/projected/b48430a2-ef1d-4747-97a4-4b6933fcea4f-kube-api-access-ggwmj\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.420703 4894 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/18cc81c1-7bef-41c2-89b7-e507624504f4-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.420712 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ff441522-39b1-47e3-be7d-d590baec8889-logs\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.420724 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/18cc81c1-7bef-41c2-89b7-e507624504f4-logs\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.420734 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/18cc81c1-7bef-41c2-89b7-e507624504f4-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.420742 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.420750 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/b48430a2-ef1d-4747-97a4-4b6933fcea4f-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.420757 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b48430a2-ef1d-4747-97a4-4b6933fcea4f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.420765 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.420773 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xm8mg\" (UniqueName: \"kubernetes.io/projected/ff441522-39b1-47e3-be7d-d590baec8889-kube-api-access-xm8mg\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.420781 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/ff441522-39b1-47e3-be7d-d590baec8889-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.420788 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p4m8h\" (UniqueName: \"kubernetes.io/projected/5c6eec80-a266-4965-9b10-c5e7b9dee04d-kube-api-access-p4m8h\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.420796 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/18cc81c1-7bef-41c2-89b7-e507624504f4-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.420804 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/ff441522-39b1-47e3-be7d-d590baec8889-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.420812 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/5c6eec80-a266-4965-9b10-c5e7b9dee04d-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.420819 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pq92x\" (UniqueName: \"kubernetes.io/projected/18cc81c1-7bef-41c2-89b7-e507624504f4-kube-api-access-pq92x\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.459687 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-56f65bfbb-tv5ld"] Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.580409 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-649c9bf67-kcmj8" event={"ID":"18cc81c1-7bef-41c2-89b7-e507624504f4","Type":"ContainerDied","Data":"eeaedfae59c5b4855abafb14f13be9032c198decab6a35bd3a3cd39ef705e2d4"} Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.580430 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-649c9bf67-kcmj8" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.588541 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-55fff446b9-mbvp9" event={"ID":"ff441522-39b1-47e3-be7d-d590baec8889","Type":"ContainerDied","Data":"2f113461bf6f89fb201c28e5cdcd43074cd02b5a622451af9af350a6df30d448"} Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.588609 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-55fff446b9-mbvp9" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.590723 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-6ff8f88b5f-gnrzd" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.593811 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.590597 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6ff8f88b5f-gnrzd" event={"ID":"1ad9eb8c-d035-456d-a345-5a4da859fc7c","Type":"ContainerDied","Data":"66c9c2c13b0b9e90ab27d0f68a9166495a171eaaf587553b4939558bd9bcbbdc"} Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.595170 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5bc56d6f79-xg28g" event={"ID":"5c6eec80-a266-4965-9b10-c5e7b9dee04d","Type":"ContainerDied","Data":"75129d3aa213adf22467d039f507f3b3892efcb2947e7979cdce3c4efe514e8b"} Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.595207 4894 scope.go:117] "RemoveContainer" containerID="bd664398729728de264c1e429a821521b44bbc925e4855a7292f30c11396b30e" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.597489 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-lfx27" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.597810 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-lfx27" event={"ID":"b48430a2-ef1d-4747-97a4-4b6933fcea4f","Type":"ContainerDied","Data":"a53ee65412c8d5fe99865aaf2cfd66727c78880d5f6b10e6d1e5f2045a2f25e1"} Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.597841 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a53ee65412c8d5fe99865aaf2cfd66727c78880d5f6b10e6d1e5f2045a2f25e1" Dec 09 15:49:32 crc kubenswrapper[4894]: E1209 15:49:32.599549 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"barbican-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-barbican-api@sha256:fe32d3ea620f0c7ecfdde9bbf28417fde03bc18c6f60b1408fa8da24d8188f16\\\"\"" pod="openstack/barbican-db-sync-s4wr5" podUID="d29fb637-ba21-4cf9-ae4d-cd9720dd5af4" Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.664738 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-55fff446b9-mbvp9"] Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.682503 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-55fff446b9-mbvp9"] Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.723757 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-649c9bf67-kcmj8"] Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.733919 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-649c9bf67-kcmj8"] Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.748535 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-6ff8f88b5f-gnrzd"] Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.754864 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-6ff8f88b5f-gnrzd"] Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.764776 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5bc56d6f79-xg28g"] Dec 09 15:49:32 crc kubenswrapper[4894]: I1209 15:49:32.772934 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5bc56d6f79-xg28g"] Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.374158 4894 scope.go:117] "RemoveContainer" containerID="eab5597ec0b22d524b0f4e7b62f0df5bc6c0595217f9c4687712ca2d09758473" Dec 09 15:49:33 crc kubenswrapper[4894]: E1209 15:49:33.402089 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b59b7445e581cc720038107e421371c86c5765b2967e77d884ef29b1d9fd0f49" Dec 09 15:49:33 crc kubenswrapper[4894]: E1209 15:49:33.402247 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b59b7445e581cc720038107e421371c86c5765b2967e77d884ef29b1d9fd0f49,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-6wwj2,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-nbnz8_openstack(ce9f3ab9-097d-4d0d-8082-08019bce834c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 15:49:33 crc kubenswrapper[4894]: E1209 15:49:33.403329 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-nbnz8" podUID="ce9f3ab9-097d-4d0d-8082-08019bce834c" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.442263 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7f69459849-bjgnn"] Dec 09 15:49:33 crc kubenswrapper[4894]: E1209 15:49:33.448763 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b48430a2-ef1d-4747-97a4-4b6933fcea4f" containerName="neutron-db-sync" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.448918 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b48430a2-ef1d-4747-97a4-4b6933fcea4f" containerName="neutron-db-sync" Dec 09 15:49:33 crc kubenswrapper[4894]: E1209 15:49:33.449021 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c6eec80-a266-4965-9b10-c5e7b9dee04d" containerName="dnsmasq-dns" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.449080 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c6eec80-a266-4965-9b10-c5e7b9dee04d" containerName="dnsmasq-dns" Dec 09 15:49:33 crc kubenswrapper[4894]: E1209 15:49:33.449174 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c6eec80-a266-4965-9b10-c5e7b9dee04d" containerName="init" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.449228 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c6eec80-a266-4965-9b10-c5e7b9dee04d" containerName="init" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.472482 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b48430a2-ef1d-4747-97a4-4b6933fcea4f" containerName="neutron-db-sync" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.472557 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c6eec80-a266-4965-9b10-c5e7b9dee04d" containerName="dnsmasq-dns" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.475327 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f69459849-bjgnn" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.478065 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f69459849-bjgnn"] Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.521467 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-56fb46ccd4-rwfjk"] Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.523023 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56fb46ccd4-rwfjk" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.530139 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-8s7s9" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.530395 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.530524 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.530700 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.539673 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-56fb46ccd4-rwfjk"] Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.569042 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-ovsdbserver-sb\") pod \"dnsmasq-dns-7f69459849-bjgnn\" (UID: \"7d66024e-378f-4433-bc75-9088982a58a9\") " pod="openstack/dnsmasq-dns-7f69459849-bjgnn" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.569132 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-dns-svc\") pod \"dnsmasq-dns-7f69459849-bjgnn\" (UID: \"7d66024e-378f-4433-bc75-9088982a58a9\") " pod="openstack/dnsmasq-dns-7f69459849-bjgnn" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.569208 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-ovsdbserver-nb\") pod \"dnsmasq-dns-7f69459849-bjgnn\" (UID: \"7d66024e-378f-4433-bc75-9088982a58a9\") " pod="openstack/dnsmasq-dns-7f69459849-bjgnn" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.569240 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tmqk6\" (UniqueName: \"kubernetes.io/projected/7d66024e-378f-4433-bc75-9088982a58a9-kube-api-access-tmqk6\") pod \"dnsmasq-dns-7f69459849-bjgnn\" (UID: \"7d66024e-378f-4433-bc75-9088982a58a9\") " pod="openstack/dnsmasq-dns-7f69459849-bjgnn" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.569526 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-config\") pod \"dnsmasq-dns-7f69459849-bjgnn\" (UID: \"7d66024e-378f-4433-bc75-9088982a58a9\") " pod="openstack/dnsmasq-dns-7f69459849-bjgnn" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.626988 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-56f65bfbb-tv5ld" event={"ID":"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16","Type":"ContainerStarted","Data":"8f1e8796e95ab32bb17dd5588ab21f261474d4ba2077138a163f5916d2f31f37"} Dec 09 15:49:33 crc kubenswrapper[4894]: E1209 15:49:33.637701 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api@sha256:b59b7445e581cc720038107e421371c86c5765b2967e77d884ef29b1d9fd0f49\\\"\"" pod="openstack/cinder-db-sync-nbnz8" podUID="ce9f3ab9-097d-4d0d-8082-08019bce834c" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.672775 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-ovsdbserver-nb\") pod \"dnsmasq-dns-7f69459849-bjgnn\" (UID: \"7d66024e-378f-4433-bc75-9088982a58a9\") " pod="openstack/dnsmasq-dns-7f69459849-bjgnn" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.672831 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tmqk6\" (UniqueName: \"kubernetes.io/projected/7d66024e-378f-4433-bc75-9088982a58a9-kube-api-access-tmqk6\") pod \"dnsmasq-dns-7f69459849-bjgnn\" (UID: \"7d66024e-378f-4433-bc75-9088982a58a9\") " pod="openstack/dnsmasq-dns-7f69459849-bjgnn" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.672883 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-config\") pod \"dnsmasq-dns-7f69459849-bjgnn\" (UID: \"7d66024e-378f-4433-bc75-9088982a58a9\") " pod="openstack/dnsmasq-dns-7f69459849-bjgnn" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.672928 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-combined-ca-bundle\") pod \"neutron-56fb46ccd4-rwfjk\" (UID: \"03372157-f0dd-4db7-8662-3b110ff31815\") " pod="openstack/neutron-56fb46ccd4-rwfjk" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.672957 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kfm2\" (UniqueName: \"kubernetes.io/projected/03372157-f0dd-4db7-8662-3b110ff31815-kube-api-access-6kfm2\") pod \"neutron-56fb46ccd4-rwfjk\" (UID: \"03372157-f0dd-4db7-8662-3b110ff31815\") " pod="openstack/neutron-56fb46ccd4-rwfjk" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.673009 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-ovsdbserver-sb\") pod \"dnsmasq-dns-7f69459849-bjgnn\" (UID: \"7d66024e-378f-4433-bc75-9088982a58a9\") " pod="openstack/dnsmasq-dns-7f69459849-bjgnn" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.673078 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-config\") pod \"neutron-56fb46ccd4-rwfjk\" (UID: \"03372157-f0dd-4db7-8662-3b110ff31815\") " pod="openstack/neutron-56fb46ccd4-rwfjk" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.673102 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-dns-svc\") pod \"dnsmasq-dns-7f69459849-bjgnn\" (UID: \"7d66024e-378f-4433-bc75-9088982a58a9\") " pod="openstack/dnsmasq-dns-7f69459849-bjgnn" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.673140 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-httpd-config\") pod \"neutron-56fb46ccd4-rwfjk\" (UID: \"03372157-f0dd-4db7-8662-3b110ff31815\") " pod="openstack/neutron-56fb46ccd4-rwfjk" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.673202 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-ovndb-tls-certs\") pod \"neutron-56fb46ccd4-rwfjk\" (UID: \"03372157-f0dd-4db7-8662-3b110ff31815\") " pod="openstack/neutron-56fb46ccd4-rwfjk" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.674014 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-ovsdbserver-nb\") pod \"dnsmasq-dns-7f69459849-bjgnn\" (UID: \"7d66024e-378f-4433-bc75-9088982a58a9\") " pod="openstack/dnsmasq-dns-7f69459849-bjgnn" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.709970 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tmqk6\" (UniqueName: \"kubernetes.io/projected/7d66024e-378f-4433-bc75-9088982a58a9-kube-api-access-tmqk6\") pod \"dnsmasq-dns-7f69459849-bjgnn\" (UID: \"7d66024e-378f-4433-bc75-9088982a58a9\") " pod="openstack/dnsmasq-dns-7f69459849-bjgnn" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.710897 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-config\") pod \"dnsmasq-dns-7f69459849-bjgnn\" (UID: \"7d66024e-378f-4433-bc75-9088982a58a9\") " pod="openstack/dnsmasq-dns-7f69459849-bjgnn" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.721853 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-ovsdbserver-sb\") pod \"dnsmasq-dns-7f69459849-bjgnn\" (UID: \"7d66024e-378f-4433-bc75-9088982a58a9\") " pod="openstack/dnsmasq-dns-7f69459849-bjgnn" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.722559 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-dns-svc\") pod \"dnsmasq-dns-7f69459849-bjgnn\" (UID: \"7d66024e-378f-4433-bc75-9088982a58a9\") " pod="openstack/dnsmasq-dns-7f69459849-bjgnn" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.774397 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-combined-ca-bundle\") pod \"neutron-56fb46ccd4-rwfjk\" (UID: \"03372157-f0dd-4db7-8662-3b110ff31815\") " pod="openstack/neutron-56fb46ccd4-rwfjk" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.774452 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kfm2\" (UniqueName: \"kubernetes.io/projected/03372157-f0dd-4db7-8662-3b110ff31815-kube-api-access-6kfm2\") pod \"neutron-56fb46ccd4-rwfjk\" (UID: \"03372157-f0dd-4db7-8662-3b110ff31815\") " pod="openstack/neutron-56fb46ccd4-rwfjk" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.774572 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-config\") pod \"neutron-56fb46ccd4-rwfjk\" (UID: \"03372157-f0dd-4db7-8662-3b110ff31815\") " pod="openstack/neutron-56fb46ccd4-rwfjk" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.774599 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-httpd-config\") pod \"neutron-56fb46ccd4-rwfjk\" (UID: \"03372157-f0dd-4db7-8662-3b110ff31815\") " pod="openstack/neutron-56fb46ccd4-rwfjk" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.774653 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-ovndb-tls-certs\") pod \"neutron-56fb46ccd4-rwfjk\" (UID: \"03372157-f0dd-4db7-8662-3b110ff31815\") " pod="openstack/neutron-56fb46ccd4-rwfjk" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.784460 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-ovndb-tls-certs\") pod \"neutron-56fb46ccd4-rwfjk\" (UID: \"03372157-f0dd-4db7-8662-3b110ff31815\") " pod="openstack/neutron-56fb46ccd4-rwfjk" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.808901 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-combined-ca-bundle\") pod \"neutron-56fb46ccd4-rwfjk\" (UID: \"03372157-f0dd-4db7-8662-3b110ff31815\") " pod="openstack/neutron-56fb46ccd4-rwfjk" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.820316 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-config\") pod \"neutron-56fb46ccd4-rwfjk\" (UID: \"03372157-f0dd-4db7-8662-3b110ff31815\") " pod="openstack/neutron-56fb46ccd4-rwfjk" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.823593 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-httpd-config\") pod \"neutron-56fb46ccd4-rwfjk\" (UID: \"03372157-f0dd-4db7-8662-3b110ff31815\") " pod="openstack/neutron-56fb46ccd4-rwfjk" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.824556 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kfm2\" (UniqueName: \"kubernetes.io/projected/03372157-f0dd-4db7-8662-3b110ff31815-kube-api-access-6kfm2\") pod \"neutron-56fb46ccd4-rwfjk\" (UID: \"03372157-f0dd-4db7-8662-3b110ff31815\") " pod="openstack/neutron-56fb46ccd4-rwfjk" Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.854910 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-rl92n"] Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.856304 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f69459849-bjgnn" Dec 09 15:49:33 crc kubenswrapper[4894]: W1209 15:49:33.872552 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfebdd308_ca10_4ba3_a34c_b000517eaeeb.slice/crio-3e44e4f53d0c69d499d9442e09b9939e6cda05f33e4aed7885f2716661f551b5 WatchSource:0}: Error finding container 3e44e4f53d0c69d499d9442e09b9939e6cda05f33e4aed7885f2716661f551b5: Status 404 returned error can't find the container with id 3e44e4f53d0c69d499d9442e09b9939e6cda05f33e4aed7885f2716661f551b5 Dec 09 15:49:33 crc kubenswrapper[4894]: I1209 15:49:33.937814 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56fb46ccd4-rwfjk" Dec 09 15:49:34 crc kubenswrapper[4894]: I1209 15:49:34.143225 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="18cc81c1-7bef-41c2-89b7-e507624504f4" path="/var/lib/kubelet/pods/18cc81c1-7bef-41c2-89b7-e507624504f4/volumes" Dec 09 15:49:34 crc kubenswrapper[4894]: I1209 15:49:34.144163 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ad9eb8c-d035-456d-a345-5a4da859fc7c" path="/var/lib/kubelet/pods/1ad9eb8c-d035-456d-a345-5a4da859fc7c/volumes" Dec 09 15:49:34 crc kubenswrapper[4894]: I1209 15:49:34.144600 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c6eec80-a266-4965-9b10-c5e7b9dee04d" path="/var/lib/kubelet/pods/5c6eec80-a266-4965-9b10-c5e7b9dee04d/volumes" Dec 09 15:49:34 crc kubenswrapper[4894]: I1209 15:49:34.145518 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ff441522-39b1-47e3-be7d-d590baec8889" path="/var/lib/kubelet/pods/ff441522-39b1-47e3-be7d-d590baec8889/volumes" Dec 09 15:49:34 crc kubenswrapper[4894]: I1209 15:49:34.146442 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizon-6b9bcbff6b-l6n7z"] Dec 09 15:49:34 crc kubenswrapper[4894]: I1209 15:49:34.401759 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7f69459849-bjgnn"] Dec 09 15:49:34 crc kubenswrapper[4894]: I1209 15:49:34.593277 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-56fb46ccd4-rwfjk"] Dec 09 15:49:34 crc kubenswrapper[4894]: I1209 15:49:34.647546 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rl92n" event={"ID":"febdd308-ca10-4ba3-a34c-b000517eaeeb","Type":"ContainerStarted","Data":"25bc8616e520df5b88b689fa0aa217ecd8bd11c7b2224dc7ce70a2f0697d7388"} Dec 09 15:49:34 crc kubenswrapper[4894]: I1209 15:49:34.648926 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rl92n" event={"ID":"febdd308-ca10-4ba3-a34c-b000517eaeeb","Type":"ContainerStarted","Data":"3e44e4f53d0c69d499d9442e09b9939e6cda05f33e4aed7885f2716661f551b5"} Dec 09 15:49:34 crc kubenswrapper[4894]: I1209 15:49:34.676384 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-rl92n" podStartSLOduration=24.676366052 podStartE2EDuration="24.676366052s" podCreationTimestamp="2025-12-09 15:49:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:49:34.674526233 +0000 UTC m=+1068.993736912" watchObservedRunningTime="2025-12-09 15:49:34.676366052 +0000 UTC m=+1068.995576721" Dec 09 15:49:34 crc kubenswrapper[4894]: I1209 15:49:34.688877 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jr2lq" event={"ID":"7d3e1277-28d7-4aa1-9121-75d859ae4688","Type":"ContainerStarted","Data":"bf7b9361fff26dd243ab8d42e89b0ce1e58793f86d9fa6fd7519b1a4d80667d7"} Dec 09 15:49:34 crc kubenswrapper[4894]: I1209 15:49:34.706373 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f69459849-bjgnn" event={"ID":"7d66024e-378f-4433-bc75-9088982a58a9","Type":"ContainerStarted","Data":"655772a72b1edb3848a5fafd5a82eb3b2bd6de7d6b6d50c6d38a0d76adc5f8c6"} Dec 09 15:49:34 crc kubenswrapper[4894]: I1209 15:49:34.706423 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f69459849-bjgnn" event={"ID":"7d66024e-378f-4433-bc75-9088982a58a9","Type":"ContainerStarted","Data":"b690b9ce22591250c9278450e01178bf9649fbde2f26bc63ec4a3c76e758e757"} Dec 09 15:49:34 crc kubenswrapper[4894]: I1209 15:49:34.713515 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-jr2lq" podStartSLOduration=4.905902201 podStartE2EDuration="35.713497032s" podCreationTimestamp="2025-12-09 15:48:59 +0000 UTC" firstStartedPulling="2025-12-09 15:49:01.235426106 +0000 UTC m=+1035.554636775" lastFinishedPulling="2025-12-09 15:49:32.043020937 +0000 UTC m=+1066.362231606" observedRunningTime="2025-12-09 15:49:34.712765793 +0000 UTC m=+1069.031976472" watchObservedRunningTime="2025-12-09 15:49:34.713497032 +0000 UTC m=+1069.032707701" Dec 09 15:49:34 crc kubenswrapper[4894]: I1209 15:49:34.715520 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bab95882-60a4-4e02-b3b7-b2a6ae55315f","Type":"ContainerStarted","Data":"42832b08996b48d3dee186127faf58ac23a8ebe1bcccbb9bc653bbc535ff663c"} Dec 09 15:49:34 crc kubenswrapper[4894]: I1209 15:49:34.717542 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56fb46ccd4-rwfjk" event={"ID":"03372157-f0dd-4db7-8662-3b110ff31815","Type":"ContainerStarted","Data":"335e3809b8720ed0c732689163dbea75739835ceae0bae80a73f7acb7c2492d9"} Dec 09 15:49:34 crc kubenswrapper[4894]: I1209 15:49:34.718361 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b9bcbff6b-l6n7z" event={"ID":"6b8de55a-230d-4dbe-9f6c-aae6138d865f","Type":"ContainerStarted","Data":"925ba1a1c5fe3eb56bfcfa70fbb6b50bf603934bf6c00e6edee405d8703971fa"} Dec 09 15:49:34 crc kubenswrapper[4894]: I1209 15:49:34.719657 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-56f65bfbb-tv5ld" event={"ID":"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16","Type":"ContainerStarted","Data":"1b3e955a174036764325c570292b5dde1384ef1df60f2868c564915e82f08177"} Dec 09 15:49:34 crc kubenswrapper[4894]: I1209 15:49:34.719687 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-56f65bfbb-tv5ld" event={"ID":"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16","Type":"ContainerStarted","Data":"b2f96b6acf368b01f8d4a3be9093d23033af84a3075e278756e462ebdc719a8b"} Dec 09 15:49:34 crc kubenswrapper[4894]: I1209 15:49:34.769453 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-56f65bfbb-tv5ld" podStartSLOduration=26.057491787 podStartE2EDuration="26.769430944s" podCreationTimestamp="2025-12-09 15:49:08 +0000 UTC" firstStartedPulling="2025-12-09 15:49:33.480901117 +0000 UTC m=+1067.800111786" lastFinishedPulling="2025-12-09 15:49:34.192840274 +0000 UTC m=+1068.512050943" observedRunningTime="2025-12-09 15:49:34.761353353 +0000 UTC m=+1069.080564022" watchObservedRunningTime="2025-12-09 15:49:34.769430944 +0000 UTC m=+1069.088641613" Dec 09 15:49:35 crc kubenswrapper[4894]: I1209 15:49:35.729125 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b9bcbff6b-l6n7z" event={"ID":"6b8de55a-230d-4dbe-9f6c-aae6138d865f","Type":"ContainerStarted","Data":"3c877218b174848d720a9534c43b79e89c8f437554dc12e4dbb3cb0d05a7ea98"} Dec 09 15:49:35 crc kubenswrapper[4894]: I1209 15:49:35.732367 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56fb46ccd4-rwfjk" event={"ID":"03372157-f0dd-4db7-8662-3b110ff31815","Type":"ContainerStarted","Data":"f1b7ca4463ac41d9e5ca80297e98942be3f547362d0beba09e8bd697366f6f4b"} Dec 09 15:49:35 crc kubenswrapper[4894]: I1209 15:49:35.732431 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56fb46ccd4-rwfjk" event={"ID":"03372157-f0dd-4db7-8662-3b110ff31815","Type":"ContainerStarted","Data":"2e721c48a84006dcd7ba5afbee0b816cbf4fc8ec61bb683872a4cac2c7e906c2"} Dec 09 15:49:35 crc kubenswrapper[4894]: I1209 15:49:35.737187 4894 generic.go:334] "Generic (PLEG): container finished" podID="7d66024e-378f-4433-bc75-9088982a58a9" containerID="655772a72b1edb3848a5fafd5a82eb3b2bd6de7d6b6d50c6d38a0d76adc5f8c6" exitCode=0 Dec 09 15:49:35 crc kubenswrapper[4894]: I1209 15:49:35.737357 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f69459849-bjgnn" event={"ID":"7d66024e-378f-4433-bc75-9088982a58a9","Type":"ContainerDied","Data":"655772a72b1edb3848a5fafd5a82eb3b2bd6de7d6b6d50c6d38a0d76adc5f8c6"} Dec 09 15:49:35 crc kubenswrapper[4894]: I1209 15:49:35.737409 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f69459849-bjgnn" event={"ID":"7d66024e-378f-4433-bc75-9088982a58a9","Type":"ContainerStarted","Data":"b7f1944245f8d45a397ef5ccdcb3ad3132508559c41047d0a95422580ee8e502"} Dec 09 15:49:35 crc kubenswrapper[4894]: I1209 15:49:35.766293 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7f69459849-bjgnn" podStartSLOduration=2.766279188 podStartE2EDuration="2.766279188s" podCreationTimestamp="2025-12-09 15:49:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:49:35.762408536 +0000 UTC m=+1070.081619205" watchObservedRunningTime="2025-12-09 15:49:35.766279188 +0000 UTC m=+1070.085489847" Dec 09 15:49:35 crc kubenswrapper[4894]: I1209 15:49:35.791247 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-bbb4b9577-djjjb"] Dec 09 15:49:35 crc kubenswrapper[4894]: I1209 15:49:35.792572 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:35 crc kubenswrapper[4894]: I1209 15:49:35.795695 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Dec 09 15:49:35 crc kubenswrapper[4894]: I1209 15:49:35.797323 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Dec 09 15:49:35 crc kubenswrapper[4894]: I1209 15:49:35.829655 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-bbb4b9577-djjjb"] Dec 09 15:49:35 crc kubenswrapper[4894]: I1209 15:49:35.927285 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/87c53f16-3692-4b98-9037-e2440a3d99fb-httpd-config\") pod \"neutron-bbb4b9577-djjjb\" (UID: \"87c53f16-3692-4b98-9037-e2440a3d99fb\") " pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:35 crc kubenswrapper[4894]: I1209 15:49:35.927587 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c53f16-3692-4b98-9037-e2440a3d99fb-public-tls-certs\") pod \"neutron-bbb4b9577-djjjb\" (UID: \"87c53f16-3692-4b98-9037-e2440a3d99fb\") " pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:35 crc kubenswrapper[4894]: I1209 15:49:35.927780 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/87c53f16-3692-4b98-9037-e2440a3d99fb-config\") pod \"neutron-bbb4b9577-djjjb\" (UID: \"87c53f16-3692-4b98-9037-e2440a3d99fb\") " pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:35 crc kubenswrapper[4894]: I1209 15:49:35.927892 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-68wpb\" (UniqueName: \"kubernetes.io/projected/87c53f16-3692-4b98-9037-e2440a3d99fb-kube-api-access-68wpb\") pod \"neutron-bbb4b9577-djjjb\" (UID: \"87c53f16-3692-4b98-9037-e2440a3d99fb\") " pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:35 crc kubenswrapper[4894]: I1209 15:49:35.927976 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c53f16-3692-4b98-9037-e2440a3d99fb-combined-ca-bundle\") pod \"neutron-bbb4b9577-djjjb\" (UID: \"87c53f16-3692-4b98-9037-e2440a3d99fb\") " pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:35 crc kubenswrapper[4894]: I1209 15:49:35.928088 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c53f16-3692-4b98-9037-e2440a3d99fb-internal-tls-certs\") pod \"neutron-bbb4b9577-djjjb\" (UID: \"87c53f16-3692-4b98-9037-e2440a3d99fb\") " pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:35 crc kubenswrapper[4894]: I1209 15:49:35.928127 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c53f16-3692-4b98-9037-e2440a3d99fb-ovndb-tls-certs\") pod \"neutron-bbb4b9577-djjjb\" (UID: \"87c53f16-3692-4b98-9037-e2440a3d99fb\") " pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:36 crc kubenswrapper[4894]: I1209 15:49:36.030387 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c53f16-3692-4b98-9037-e2440a3d99fb-ovndb-tls-certs\") pod \"neutron-bbb4b9577-djjjb\" (UID: \"87c53f16-3692-4b98-9037-e2440a3d99fb\") " pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:36 crc kubenswrapper[4894]: I1209 15:49:36.030726 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/87c53f16-3692-4b98-9037-e2440a3d99fb-httpd-config\") pod \"neutron-bbb4b9577-djjjb\" (UID: \"87c53f16-3692-4b98-9037-e2440a3d99fb\") " pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:36 crc kubenswrapper[4894]: I1209 15:49:36.030766 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c53f16-3692-4b98-9037-e2440a3d99fb-public-tls-certs\") pod \"neutron-bbb4b9577-djjjb\" (UID: \"87c53f16-3692-4b98-9037-e2440a3d99fb\") " pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:36 crc kubenswrapper[4894]: I1209 15:49:36.030819 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/87c53f16-3692-4b98-9037-e2440a3d99fb-config\") pod \"neutron-bbb4b9577-djjjb\" (UID: \"87c53f16-3692-4b98-9037-e2440a3d99fb\") " pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:36 crc kubenswrapper[4894]: I1209 15:49:36.030851 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-68wpb\" (UniqueName: \"kubernetes.io/projected/87c53f16-3692-4b98-9037-e2440a3d99fb-kube-api-access-68wpb\") pod \"neutron-bbb4b9577-djjjb\" (UID: \"87c53f16-3692-4b98-9037-e2440a3d99fb\") " pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:36 crc kubenswrapper[4894]: I1209 15:49:36.030874 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c53f16-3692-4b98-9037-e2440a3d99fb-combined-ca-bundle\") pod \"neutron-bbb4b9577-djjjb\" (UID: \"87c53f16-3692-4b98-9037-e2440a3d99fb\") " pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:36 crc kubenswrapper[4894]: I1209 15:49:36.030978 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c53f16-3692-4b98-9037-e2440a3d99fb-internal-tls-certs\") pod \"neutron-bbb4b9577-djjjb\" (UID: \"87c53f16-3692-4b98-9037-e2440a3d99fb\") " pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:36 crc kubenswrapper[4894]: I1209 15:49:36.039375 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/87c53f16-3692-4b98-9037-e2440a3d99fb-httpd-config\") pod \"neutron-bbb4b9577-djjjb\" (UID: \"87c53f16-3692-4b98-9037-e2440a3d99fb\") " pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:36 crc kubenswrapper[4894]: I1209 15:49:36.039411 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c53f16-3692-4b98-9037-e2440a3d99fb-internal-tls-certs\") pod \"neutron-bbb4b9577-djjjb\" (UID: \"87c53f16-3692-4b98-9037-e2440a3d99fb\") " pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:36 crc kubenswrapper[4894]: I1209 15:49:36.039696 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c53f16-3692-4b98-9037-e2440a3d99fb-ovndb-tls-certs\") pod \"neutron-bbb4b9577-djjjb\" (UID: \"87c53f16-3692-4b98-9037-e2440a3d99fb\") " pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:36 crc kubenswrapper[4894]: I1209 15:49:36.040449 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/87c53f16-3692-4b98-9037-e2440a3d99fb-config\") pod \"neutron-bbb4b9577-djjjb\" (UID: \"87c53f16-3692-4b98-9037-e2440a3d99fb\") " pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:36 crc kubenswrapper[4894]: I1209 15:49:36.041089 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/87c53f16-3692-4b98-9037-e2440a3d99fb-combined-ca-bundle\") pod \"neutron-bbb4b9577-djjjb\" (UID: \"87c53f16-3692-4b98-9037-e2440a3d99fb\") " pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:36 crc kubenswrapper[4894]: I1209 15:49:36.041402 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/87c53f16-3692-4b98-9037-e2440a3d99fb-public-tls-certs\") pod \"neutron-bbb4b9577-djjjb\" (UID: \"87c53f16-3692-4b98-9037-e2440a3d99fb\") " pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:36 crc kubenswrapper[4894]: I1209 15:49:36.064458 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-68wpb\" (UniqueName: \"kubernetes.io/projected/87c53f16-3692-4b98-9037-e2440a3d99fb-kube-api-access-68wpb\") pod \"neutron-bbb4b9577-djjjb\" (UID: \"87c53f16-3692-4b98-9037-e2440a3d99fb\") " pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:36 crc kubenswrapper[4894]: I1209 15:49:36.113108 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:36 crc kubenswrapper[4894]: I1209 15:49:36.747568 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-6b9bcbff6b-l6n7z" event={"ID":"6b8de55a-230d-4dbe-9f6c-aae6138d865f","Type":"ContainerStarted","Data":"0934420508c5522954250f153be2b513896c1c06790512d431d065afb883e28d"} Dec 09 15:49:36 crc kubenswrapper[4894]: I1209 15:49:36.748071 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7f69459849-bjgnn" Dec 09 15:49:36 crc kubenswrapper[4894]: I1209 15:49:36.772889 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-56fb46ccd4-rwfjk" podStartSLOduration=3.772871596 podStartE2EDuration="3.772871596s" podCreationTimestamp="2025-12-09 15:49:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:49:36.7707044 +0000 UTC m=+1071.089915079" watchObservedRunningTime="2025-12-09 15:49:36.772871596 +0000 UTC m=+1071.092082265" Dec 09 15:49:36 crc kubenswrapper[4894]: I1209 15:49:36.793474 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizon-6b9bcbff6b-l6n7z" podStartSLOduration=28.311900008 podStartE2EDuration="28.793453534s" podCreationTimestamp="2025-12-09 15:49:08 +0000 UTC" firstStartedPulling="2025-12-09 15:49:34.157386488 +0000 UTC m=+1068.476597157" lastFinishedPulling="2025-12-09 15:49:34.638940014 +0000 UTC m=+1068.958150683" observedRunningTime="2025-12-09 15:49:36.790502367 +0000 UTC m=+1071.109713036" watchObservedRunningTime="2025-12-09 15:49:36.793453534 +0000 UTC m=+1071.112664203" Dec 09 15:49:37 crc kubenswrapper[4894]: I1209 15:49:37.830754 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-bbb4b9577-djjjb"] Dec 09 15:49:38 crc kubenswrapper[4894]: I1209 15:49:38.563954 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:38 crc kubenswrapper[4894]: I1209 15:49:38.565343 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:49:38 crc kubenswrapper[4894]: I1209 15:49:38.691998 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:38 crc kubenswrapper[4894]: I1209 15:49:38.692051 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:49:38 crc kubenswrapper[4894]: I1209 15:49:38.773165 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bbb4b9577-djjjb" event={"ID":"87c53f16-3692-4b98-9037-e2440a3d99fb","Type":"ContainerStarted","Data":"a884411f9ca5dc185802301e921a1bedd3761d3aef6a07d79c0750859ead5ef0"} Dec 09 15:49:38 crc kubenswrapper[4894]: I1209 15:49:38.773606 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bbb4b9577-djjjb" event={"ID":"87c53f16-3692-4b98-9037-e2440a3d99fb","Type":"ContainerStarted","Data":"4c12d163f28fffff1c7da2d036b40b71a3f97568390d6c5f1a344db9ede7d256"} Dec 09 15:49:38 crc kubenswrapper[4894]: I1209 15:49:38.773622 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-bbb4b9577-djjjb" event={"ID":"87c53f16-3692-4b98-9037-e2440a3d99fb","Type":"ContainerStarted","Data":"807542a396a986ec7445befdccf3d9e803d0ee46b03047c2d3d5092a2cf3867c"} Dec 09 15:49:38 crc kubenswrapper[4894]: I1209 15:49:38.773700 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:49:38 crc kubenswrapper[4894]: I1209 15:49:38.777395 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bab95882-60a4-4e02-b3b7-b2a6ae55315f","Type":"ContainerStarted","Data":"ad993b83ca0bae3b9d2a6aa998c2747bd8ca7896fb6279c3e80ae74e1c99bd1f"} Dec 09 15:49:38 crc kubenswrapper[4894]: I1209 15:49:38.779510 4894 generic.go:334] "Generic (PLEG): container finished" podID="7d3e1277-28d7-4aa1-9121-75d859ae4688" containerID="bf7b9361fff26dd243ab8d42e89b0ce1e58793f86d9fa6fd7519b1a4d80667d7" exitCode=0 Dec 09 15:49:38 crc kubenswrapper[4894]: I1209 15:49:38.780842 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jr2lq" event={"ID":"7d3e1277-28d7-4aa1-9121-75d859ae4688","Type":"ContainerDied","Data":"bf7b9361fff26dd243ab8d42e89b0ce1e58793f86d9fa6fd7519b1a4d80667d7"} Dec 09 15:49:38 crc kubenswrapper[4894]: I1209 15:49:38.807152 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-bbb4b9577-djjjb" podStartSLOduration=3.807132143 podStartE2EDuration="3.807132143s" podCreationTimestamp="2025-12-09 15:49:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:49:38.801725551 +0000 UTC m=+1073.120936210" watchObservedRunningTime="2025-12-09 15:49:38.807132143 +0000 UTC m=+1073.126342802" Dec 09 15:49:39 crc kubenswrapper[4894]: I1209 15:49:39.789841 4894 generic.go:334] "Generic (PLEG): container finished" podID="febdd308-ca10-4ba3-a34c-b000517eaeeb" containerID="25bc8616e520df5b88b689fa0aa217ecd8bd11c7b2224dc7ce70a2f0697d7388" exitCode=0 Dec 09 15:49:39 crc kubenswrapper[4894]: I1209 15:49:39.789897 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rl92n" event={"ID":"febdd308-ca10-4ba3-a34c-b000517eaeeb","Type":"ContainerDied","Data":"25bc8616e520df5b88b689fa0aa217ecd8bd11c7b2224dc7ce70a2f0697d7388"} Dec 09 15:49:40 crc kubenswrapper[4894]: I1209 15:49:40.140000 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jr2lq" Dec 09 15:49:40 crc kubenswrapper[4894]: I1209 15:49:40.241703 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d3e1277-28d7-4aa1-9121-75d859ae4688-logs\") pod \"7d3e1277-28d7-4aa1-9121-75d859ae4688\" (UID: \"7d3e1277-28d7-4aa1-9121-75d859ae4688\") " Dec 09 15:49:40 crc kubenswrapper[4894]: I1209 15:49:40.241752 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d3e1277-28d7-4aa1-9121-75d859ae4688-scripts\") pod \"7d3e1277-28d7-4aa1-9121-75d859ae4688\" (UID: \"7d3e1277-28d7-4aa1-9121-75d859ae4688\") " Dec 09 15:49:40 crc kubenswrapper[4894]: I1209 15:49:40.241784 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d3e1277-28d7-4aa1-9121-75d859ae4688-combined-ca-bundle\") pod \"7d3e1277-28d7-4aa1-9121-75d859ae4688\" (UID: \"7d3e1277-28d7-4aa1-9121-75d859ae4688\") " Dec 09 15:49:40 crc kubenswrapper[4894]: I1209 15:49:40.241858 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d3e1277-28d7-4aa1-9121-75d859ae4688-config-data\") pod \"7d3e1277-28d7-4aa1-9121-75d859ae4688\" (UID: \"7d3e1277-28d7-4aa1-9121-75d859ae4688\") " Dec 09 15:49:40 crc kubenswrapper[4894]: I1209 15:49:40.241936 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bc5tl\" (UniqueName: \"kubernetes.io/projected/7d3e1277-28d7-4aa1-9121-75d859ae4688-kube-api-access-bc5tl\") pod \"7d3e1277-28d7-4aa1-9121-75d859ae4688\" (UID: \"7d3e1277-28d7-4aa1-9121-75d859ae4688\") " Dec 09 15:49:40 crc kubenswrapper[4894]: I1209 15:49:40.242943 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7d3e1277-28d7-4aa1-9121-75d859ae4688-logs" (OuterVolumeSpecName: "logs") pod "7d3e1277-28d7-4aa1-9121-75d859ae4688" (UID: "7d3e1277-28d7-4aa1-9121-75d859ae4688"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:49:40 crc kubenswrapper[4894]: I1209 15:49:40.248431 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d3e1277-28d7-4aa1-9121-75d859ae4688-scripts" (OuterVolumeSpecName: "scripts") pod "7d3e1277-28d7-4aa1-9121-75d859ae4688" (UID: "7d3e1277-28d7-4aa1-9121-75d859ae4688"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:40 crc kubenswrapper[4894]: I1209 15:49:40.249136 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d3e1277-28d7-4aa1-9121-75d859ae4688-kube-api-access-bc5tl" (OuterVolumeSpecName: "kube-api-access-bc5tl") pod "7d3e1277-28d7-4aa1-9121-75d859ae4688" (UID: "7d3e1277-28d7-4aa1-9121-75d859ae4688"). InnerVolumeSpecName "kube-api-access-bc5tl". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:49:40 crc kubenswrapper[4894]: I1209 15:49:40.279807 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d3e1277-28d7-4aa1-9121-75d859ae4688-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7d3e1277-28d7-4aa1-9121-75d859ae4688" (UID: "7d3e1277-28d7-4aa1-9121-75d859ae4688"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:40 crc kubenswrapper[4894]: I1209 15:49:40.279851 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7d3e1277-28d7-4aa1-9121-75d859ae4688-config-data" (OuterVolumeSpecName: "config-data") pod "7d3e1277-28d7-4aa1-9121-75d859ae4688" (UID: "7d3e1277-28d7-4aa1-9121-75d859ae4688"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:40 crc kubenswrapper[4894]: I1209 15:49:40.343945 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bc5tl\" (UniqueName: \"kubernetes.io/projected/7d3e1277-28d7-4aa1-9121-75d859ae4688-kube-api-access-bc5tl\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:40 crc kubenswrapper[4894]: I1209 15:49:40.343988 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7d3e1277-28d7-4aa1-9121-75d859ae4688-logs\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:40 crc kubenswrapper[4894]: I1209 15:49:40.344004 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7d3e1277-28d7-4aa1-9121-75d859ae4688-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:40 crc kubenswrapper[4894]: I1209 15:49:40.344016 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7d3e1277-28d7-4aa1-9121-75d859ae4688-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:40 crc kubenswrapper[4894]: I1209 15:49:40.344028 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7d3e1277-28d7-4aa1-9121-75d859ae4688-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:40 crc kubenswrapper[4894]: I1209 15:49:40.799935 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-jr2lq" event={"ID":"7d3e1277-28d7-4aa1-9121-75d859ae4688","Type":"ContainerDied","Data":"79c97c797089c4b3b8dc88a13bacdb638d7e52432f17a350ea19b191336ead57"} Dec 09 15:49:40 crc kubenswrapper[4894]: I1209 15:49:40.799981 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="79c97c797089c4b3b8dc88a13bacdb638d7e52432f17a350ea19b191336ead57" Dec 09 15:49:40 crc kubenswrapper[4894]: I1209 15:49:40.799952 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-jr2lq" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.000447 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-59f4fb4758-w8ndv"] Dec 09 15:49:41 crc kubenswrapper[4894]: E1209 15:49:41.001043 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d3e1277-28d7-4aa1-9121-75d859ae4688" containerName="placement-db-sync" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.001062 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d3e1277-28d7-4aa1-9121-75d859ae4688" containerName="placement-db-sync" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.001327 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d3e1277-28d7-4aa1-9121-75d859ae4688" containerName="placement-db-sync" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.003039 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.023444 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.023516 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.023831 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.023933 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-bjd4l" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.024873 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.044279 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-59f4fb4758-w8ndv"] Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.163349 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10ff6ecf-af86-4595-bff7-ad0fd49eb3e1-scripts\") pod \"placement-59f4fb4758-w8ndv\" (UID: \"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1\") " pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.163464 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10ff6ecf-af86-4595-bff7-ad0fd49eb3e1-combined-ca-bundle\") pod \"placement-59f4fb4758-w8ndv\" (UID: \"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1\") " pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.163539 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gnj26\" (UniqueName: \"kubernetes.io/projected/10ff6ecf-af86-4595-bff7-ad0fd49eb3e1-kube-api-access-gnj26\") pod \"placement-59f4fb4758-w8ndv\" (UID: \"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1\") " pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.163594 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10ff6ecf-af86-4595-bff7-ad0fd49eb3e1-config-data\") pod \"placement-59f4fb4758-w8ndv\" (UID: \"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1\") " pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.163654 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/10ff6ecf-af86-4595-bff7-ad0fd49eb3e1-internal-tls-certs\") pod \"placement-59f4fb4758-w8ndv\" (UID: \"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1\") " pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.164762 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/10ff6ecf-af86-4595-bff7-ad0fd49eb3e1-public-tls-certs\") pod \"placement-59f4fb4758-w8ndv\" (UID: \"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1\") " pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.164805 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10ff6ecf-af86-4595-bff7-ad0fd49eb3e1-logs\") pod \"placement-59f4fb4758-w8ndv\" (UID: \"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1\") " pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.269124 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10ff6ecf-af86-4595-bff7-ad0fd49eb3e1-scripts\") pod \"placement-59f4fb4758-w8ndv\" (UID: \"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1\") " pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.269166 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10ff6ecf-af86-4595-bff7-ad0fd49eb3e1-combined-ca-bundle\") pod \"placement-59f4fb4758-w8ndv\" (UID: \"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1\") " pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.269206 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gnj26\" (UniqueName: \"kubernetes.io/projected/10ff6ecf-af86-4595-bff7-ad0fd49eb3e1-kube-api-access-gnj26\") pod \"placement-59f4fb4758-w8ndv\" (UID: \"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1\") " pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.269227 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10ff6ecf-af86-4595-bff7-ad0fd49eb3e1-config-data\") pod \"placement-59f4fb4758-w8ndv\" (UID: \"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1\") " pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.269257 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/10ff6ecf-af86-4595-bff7-ad0fd49eb3e1-internal-tls-certs\") pod \"placement-59f4fb4758-w8ndv\" (UID: \"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1\") " pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.269302 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/10ff6ecf-af86-4595-bff7-ad0fd49eb3e1-public-tls-certs\") pod \"placement-59f4fb4758-w8ndv\" (UID: \"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1\") " pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.269316 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10ff6ecf-af86-4595-bff7-ad0fd49eb3e1-logs\") pod \"placement-59f4fb4758-w8ndv\" (UID: \"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1\") " pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.269736 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/10ff6ecf-af86-4595-bff7-ad0fd49eb3e1-logs\") pod \"placement-59f4fb4758-w8ndv\" (UID: \"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1\") " pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.276246 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/10ff6ecf-af86-4595-bff7-ad0fd49eb3e1-public-tls-certs\") pod \"placement-59f4fb4758-w8ndv\" (UID: \"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1\") " pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.276348 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/10ff6ecf-af86-4595-bff7-ad0fd49eb3e1-internal-tls-certs\") pod \"placement-59f4fb4758-w8ndv\" (UID: \"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1\") " pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.279852 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/10ff6ecf-af86-4595-bff7-ad0fd49eb3e1-scripts\") pod \"placement-59f4fb4758-w8ndv\" (UID: \"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1\") " pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.282607 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/10ff6ecf-af86-4595-bff7-ad0fd49eb3e1-config-data\") pod \"placement-59f4fb4758-w8ndv\" (UID: \"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1\") " pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.285350 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10ff6ecf-af86-4595-bff7-ad0fd49eb3e1-combined-ca-bundle\") pod \"placement-59f4fb4758-w8ndv\" (UID: \"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1\") " pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.289084 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gnj26\" (UniqueName: \"kubernetes.io/projected/10ff6ecf-af86-4595-bff7-ad0fd49eb3e1-kube-api-access-gnj26\") pod \"placement-59f4fb4758-w8ndv\" (UID: \"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1\") " pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:41 crc kubenswrapper[4894]: I1209 15:49:41.357234 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.257873 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.258185 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.270853 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rl92n" Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.288917 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-config-data\") pod \"febdd308-ca10-4ba3-a34c-b000517eaeeb\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.289025 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-credential-keys\") pod \"febdd308-ca10-4ba3-a34c-b000517eaeeb\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.289072 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mlb5l\" (UniqueName: \"kubernetes.io/projected/febdd308-ca10-4ba3-a34c-b000517eaeeb-kube-api-access-mlb5l\") pod \"febdd308-ca10-4ba3-a34c-b000517eaeeb\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.289118 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-combined-ca-bundle\") pod \"febdd308-ca10-4ba3-a34c-b000517eaeeb\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.289160 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-scripts\") pod \"febdd308-ca10-4ba3-a34c-b000517eaeeb\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.289200 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-fernet-keys\") pod \"febdd308-ca10-4ba3-a34c-b000517eaeeb\" (UID: \"febdd308-ca10-4ba3-a34c-b000517eaeeb\") " Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.294407 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/febdd308-ca10-4ba3-a34c-b000517eaeeb-kube-api-access-mlb5l" (OuterVolumeSpecName: "kube-api-access-mlb5l") pod "febdd308-ca10-4ba3-a34c-b000517eaeeb" (UID: "febdd308-ca10-4ba3-a34c-b000517eaeeb"). InnerVolumeSpecName "kube-api-access-mlb5l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.296378 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "febdd308-ca10-4ba3-a34c-b000517eaeeb" (UID: "febdd308-ca10-4ba3-a34c-b000517eaeeb"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.297084 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "febdd308-ca10-4ba3-a34c-b000517eaeeb" (UID: "febdd308-ca10-4ba3-a34c-b000517eaeeb"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.307036 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-scripts" (OuterVolumeSpecName: "scripts") pod "febdd308-ca10-4ba3-a34c-b000517eaeeb" (UID: "febdd308-ca10-4ba3-a34c-b000517eaeeb"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.319885 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-config-data" (OuterVolumeSpecName: "config-data") pod "febdd308-ca10-4ba3-a34c-b000517eaeeb" (UID: "febdd308-ca10-4ba3-a34c-b000517eaeeb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.345809 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "febdd308-ca10-4ba3-a34c-b000517eaeeb" (UID: "febdd308-ca10-4ba3-a34c-b000517eaeeb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.391619 4894 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-credential-keys\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.391686 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mlb5l\" (UniqueName: \"kubernetes.io/projected/febdd308-ca10-4ba3-a34c-b000517eaeeb-kube-api-access-mlb5l\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.391701 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.391716 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.391728 4894 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.391737 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/febdd308-ca10-4ba3-a34c-b000517eaeeb-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.816132 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-rl92n" event={"ID":"febdd308-ca10-4ba3-a34c-b000517eaeeb","Type":"ContainerDied","Data":"3e44e4f53d0c69d499d9442e09b9939e6cda05f33e4aed7885f2716661f551b5"} Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.816174 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3e44e4f53d0c69d499d9442e09b9939e6cda05f33e4aed7885f2716661f551b5" Dec 09 15:49:42 crc kubenswrapper[4894]: I1209 15:49:42.816200 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-rl92n" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.355188 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-ccc8cf6fb-9hvcd"] Dec 09 15:49:43 crc kubenswrapper[4894]: E1209 15:49:43.355738 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="febdd308-ca10-4ba3-a34c-b000517eaeeb" containerName="keystone-bootstrap" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.355751 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="febdd308-ca10-4ba3-a34c-b000517eaeeb" containerName="keystone-bootstrap" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.355923 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="febdd308-ca10-4ba3-a34c-b000517eaeeb" containerName="keystone-bootstrap" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.356445 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.360208 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.360446 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.360699 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.360843 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.361023 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.365240 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-g8fmh" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.367927 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-ccc8cf6fb-9hvcd"] Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.410430 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f70e1d59-2f1f-4d88-ac43-dda028386a96-internal-tls-certs\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.410541 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f70e1d59-2f1f-4d88-ac43-dda028386a96-credential-keys\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.410599 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f70e1d59-2f1f-4d88-ac43-dda028386a96-public-tls-certs\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.410669 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wv5hm\" (UniqueName: \"kubernetes.io/projected/f70e1d59-2f1f-4d88-ac43-dda028386a96-kube-api-access-wv5hm\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.410807 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f70e1d59-2f1f-4d88-ac43-dda028386a96-scripts\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.410875 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f70e1d59-2f1f-4d88-ac43-dda028386a96-combined-ca-bundle\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.410921 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f70e1d59-2f1f-4d88-ac43-dda028386a96-config-data\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.411110 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f70e1d59-2f1f-4d88-ac43-dda028386a96-fernet-keys\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.512896 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f70e1d59-2f1f-4d88-ac43-dda028386a96-config-data\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.512971 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f70e1d59-2f1f-4d88-ac43-dda028386a96-fernet-keys\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.513022 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f70e1d59-2f1f-4d88-ac43-dda028386a96-internal-tls-certs\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.513098 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f70e1d59-2f1f-4d88-ac43-dda028386a96-credential-keys\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.513145 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f70e1d59-2f1f-4d88-ac43-dda028386a96-public-tls-certs\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.513174 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wv5hm\" (UniqueName: \"kubernetes.io/projected/f70e1d59-2f1f-4d88-ac43-dda028386a96-kube-api-access-wv5hm\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.513222 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f70e1d59-2f1f-4d88-ac43-dda028386a96-scripts\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.513249 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f70e1d59-2f1f-4d88-ac43-dda028386a96-combined-ca-bundle\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.518292 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/f70e1d59-2f1f-4d88-ac43-dda028386a96-credential-keys\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.520066 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/f70e1d59-2f1f-4d88-ac43-dda028386a96-fernet-keys\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.520787 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f70e1d59-2f1f-4d88-ac43-dda028386a96-public-tls-certs\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.522219 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f70e1d59-2f1f-4d88-ac43-dda028386a96-combined-ca-bundle\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.530749 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f70e1d59-2f1f-4d88-ac43-dda028386a96-internal-tls-certs\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.533992 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f70e1d59-2f1f-4d88-ac43-dda028386a96-scripts\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.534907 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wv5hm\" (UniqueName: \"kubernetes.io/projected/f70e1d59-2f1f-4d88-ac43-dda028386a96-kube-api-access-wv5hm\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.540238 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f70e1d59-2f1f-4d88-ac43-dda028386a96-config-data\") pod \"keystone-ccc8cf6fb-9hvcd\" (UID: \"f70e1d59-2f1f-4d88-ac43-dda028386a96\") " pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.673324 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.859560 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7f69459849-bjgnn" Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.926522 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-694dbb6647-fzs9q"] Dec 09 15:49:43 crc kubenswrapper[4894]: I1209 15:49:43.926811 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" podUID="f20c656d-4752-41ce-a1aa-aa6980430e90" containerName="dnsmasq-dns" containerID="cri-o://09c02c60b86e32b44013a0492a40353ce279985637cd09e8deeb9bbcfac8cbbb" gracePeriod=10 Dec 09 15:49:44 crc kubenswrapper[4894]: I1209 15:49:44.854422 4894 generic.go:334] "Generic (PLEG): container finished" podID="f20c656d-4752-41ce-a1aa-aa6980430e90" containerID="09c02c60b86e32b44013a0492a40353ce279985637cd09e8deeb9bbcfac8cbbb" exitCode=0 Dec 09 15:49:44 crc kubenswrapper[4894]: I1209 15:49:44.854686 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" event={"ID":"f20c656d-4752-41ce-a1aa-aa6980430e90","Type":"ContainerDied","Data":"09c02c60b86e32b44013a0492a40353ce279985637cd09e8deeb9bbcfac8cbbb"} Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.023542 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.035950 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qc5nm\" (UniqueName: \"kubernetes.io/projected/f20c656d-4752-41ce-a1aa-aa6980430e90-kube-api-access-qc5nm\") pod \"f20c656d-4752-41ce-a1aa-aa6980430e90\" (UID: \"f20c656d-4752-41ce-a1aa-aa6980430e90\") " Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.036039 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-config\") pod \"f20c656d-4752-41ce-a1aa-aa6980430e90\" (UID: \"f20c656d-4752-41ce-a1aa-aa6980430e90\") " Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.036075 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-ovsdbserver-nb\") pod \"f20c656d-4752-41ce-a1aa-aa6980430e90\" (UID: \"f20c656d-4752-41ce-a1aa-aa6980430e90\") " Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.036175 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-dns-svc\") pod \"f20c656d-4752-41ce-a1aa-aa6980430e90\" (UID: \"f20c656d-4752-41ce-a1aa-aa6980430e90\") " Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.036227 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-ovsdbserver-sb\") pod \"f20c656d-4752-41ce-a1aa-aa6980430e90\" (UID: \"f20c656d-4752-41ce-a1aa-aa6980430e90\") " Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.045309 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f20c656d-4752-41ce-a1aa-aa6980430e90-kube-api-access-qc5nm" (OuterVolumeSpecName: "kube-api-access-qc5nm") pod "f20c656d-4752-41ce-a1aa-aa6980430e90" (UID: "f20c656d-4752-41ce-a1aa-aa6980430e90"). InnerVolumeSpecName "kube-api-access-qc5nm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.138353 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qc5nm\" (UniqueName: \"kubernetes.io/projected/f20c656d-4752-41ce-a1aa-aa6980430e90-kube-api-access-qc5nm\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.176620 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f20c656d-4752-41ce-a1aa-aa6980430e90" (UID: "f20c656d-4752-41ce-a1aa-aa6980430e90"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.194356 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f20c656d-4752-41ce-a1aa-aa6980430e90" (UID: "f20c656d-4752-41ce-a1aa-aa6980430e90"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.199601 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f20c656d-4752-41ce-a1aa-aa6980430e90" (UID: "f20c656d-4752-41ce-a1aa-aa6980430e90"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.201080 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-config" (OuterVolumeSpecName: "config") pod "f20c656d-4752-41ce-a1aa-aa6980430e90" (UID: "f20c656d-4752-41ce-a1aa-aa6980430e90"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.240420 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.240480 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.240499 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.240512 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f20c656d-4752-41ce-a1aa-aa6980430e90-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.256725 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-59f4fb4758-w8ndv"] Dec 09 15:49:45 crc kubenswrapper[4894]: W1209 15:49:45.263970 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod10ff6ecf_af86_4595_bff7_ad0fd49eb3e1.slice/crio-34f2a41ed2059a38b6e209da4172f954d125892ed43a4c6cf9452fa3fcfb0ec0 WatchSource:0}: Error finding container 34f2a41ed2059a38b6e209da4172f954d125892ed43a4c6cf9452fa3fcfb0ec0: Status 404 returned error can't find the container with id 34f2a41ed2059a38b6e209da4172f954d125892ed43a4c6cf9452fa3fcfb0ec0 Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.353187 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-ccc8cf6fb-9hvcd"] Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.876086 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bab95882-60a4-4e02-b3b7-b2a6ae55315f","Type":"ContainerStarted","Data":"36c3822c1e6a90d261e05cba628235eab698bf1a47839de9a35bce0fe9dc2b06"} Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.881692 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" event={"ID":"f20c656d-4752-41ce-a1aa-aa6980430e90","Type":"ContainerDied","Data":"0690d6aa19b142e8028f2fb342968df9a4acb8bcbb77a1d1c891166839c8b990"} Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.881766 4894 scope.go:117] "RemoveContainer" containerID="09c02c60b86e32b44013a0492a40353ce279985637cd09e8deeb9bbcfac8cbbb" Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.881927 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-694dbb6647-fzs9q" Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.889395 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ccc8cf6fb-9hvcd" event={"ID":"f70e1d59-2f1f-4d88-ac43-dda028386a96","Type":"ContainerStarted","Data":"8f4fe6bcc0da41c21bd953f9819fa4a5d3fc078431ecf4ee9b656099cf0112f0"} Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.889443 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-ccc8cf6fb-9hvcd" event={"ID":"f70e1d59-2f1f-4d88-ac43-dda028386a96","Type":"ContainerStarted","Data":"893c9bc7d0d459d5de24607699c7a4d262f01329e1f3f1b32371f9017487aa04"} Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.890870 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.894091 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-59f4fb4758-w8ndv" event={"ID":"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1","Type":"ContainerStarted","Data":"7e0f71b4029d9d228b68f87a343735473d892cbb6eba44d48109d33e4876476f"} Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.894128 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-59f4fb4758-w8ndv" event={"ID":"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1","Type":"ContainerStarted","Data":"2f9fcac877310e38745945718ad6ab0ba779b2efb488bf1e14b57fbca48ea259"} Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.894138 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-59f4fb4758-w8ndv" event={"ID":"10ff6ecf-af86-4595-bff7-ad0fd49eb3e1","Type":"ContainerStarted","Data":"34f2a41ed2059a38b6e209da4172f954d125892ed43a4c6cf9452fa3fcfb0ec0"} Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.894727 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.894757 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.920832 4894 scope.go:117] "RemoveContainer" containerID="9fbc25fcc1a3e6babc8975e83d644bc14327478c515a5f66a069e085fa1776b0" Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.924278 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-ccc8cf6fb-9hvcd" podStartSLOduration=2.9242636749999997 podStartE2EDuration="2.924263675s" podCreationTimestamp="2025-12-09 15:49:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:49:45.907138227 +0000 UTC m=+1080.226348916" watchObservedRunningTime="2025-12-09 15:49:45.924263675 +0000 UTC m=+1080.243474344" Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.935672 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-694dbb6647-fzs9q"] Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.946380 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-694dbb6647-fzs9q"] Dec 09 15:49:45 crc kubenswrapper[4894]: I1209 15:49:45.958024 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-59f4fb4758-w8ndv" podStartSLOduration=5.958006657 podStartE2EDuration="5.958006657s" podCreationTimestamp="2025-12-09 15:49:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:49:45.953421997 +0000 UTC m=+1080.272632666" watchObservedRunningTime="2025-12-09 15:49:45.958006657 +0000 UTC m=+1080.277217316" Dec 09 15:49:46 crc kubenswrapper[4894]: I1209 15:49:46.123003 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f20c656d-4752-41ce-a1aa-aa6980430e90" path="/var/lib/kubelet/pods/f20c656d-4752-41ce-a1aa-aa6980430e90/volumes" Dec 09 15:49:46 crc kubenswrapper[4894]: I1209 15:49:46.904968 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-nbnz8" event={"ID":"ce9f3ab9-097d-4d0d-8082-08019bce834c","Type":"ContainerStarted","Data":"f2e5fa88e421d711a8f091021fdd47f49dfbfe3a8d6e38495324d1020a7b71ea"} Dec 09 15:49:46 crc kubenswrapper[4894]: I1209 15:49:46.924988 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-nbnz8" podStartSLOduration=3.402470047 podStartE2EDuration="47.924938279s" podCreationTimestamp="2025-12-09 15:48:59 +0000 UTC" firstStartedPulling="2025-12-09 15:49:01.12389155 +0000 UTC m=+1035.443102219" lastFinishedPulling="2025-12-09 15:49:45.646359782 +0000 UTC m=+1079.965570451" observedRunningTime="2025-12-09 15:49:46.920241135 +0000 UTC m=+1081.239451804" watchObservedRunningTime="2025-12-09 15:49:46.924938279 +0000 UTC m=+1081.244148948" Dec 09 15:49:47 crc kubenswrapper[4894]: I1209 15:49:47.918988 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-s4wr5" event={"ID":"d29fb637-ba21-4cf9-ae4d-cd9720dd5af4","Type":"ContainerStarted","Data":"7a1928ef8d06e25be6e0fab76f6d003eaa7e0f5678729dc4598fa1ebcfbe1d67"} Dec 09 15:49:47 crc kubenswrapper[4894]: I1209 15:49:47.938588 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-s4wr5" podStartSLOduration=2.404543385 podStartE2EDuration="48.938570741s" podCreationTimestamp="2025-12-09 15:48:59 +0000 UTC" firstStartedPulling="2025-12-09 15:49:01.170523479 +0000 UTC m=+1035.489734148" lastFinishedPulling="2025-12-09 15:49:47.704550835 +0000 UTC m=+1082.023761504" observedRunningTime="2025-12-09 15:49:47.934085114 +0000 UTC m=+1082.253295803" watchObservedRunningTime="2025-12-09 15:49:47.938570741 +0000 UTC m=+1082.257781410" Dec 09 15:49:48 crc kubenswrapper[4894]: I1209 15:49:48.567454 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-56f65bfbb-tv5ld" podUID="0e41ac7e-15c8-4a9e-ae96-2e68b055ea16" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 09 15:49:48 crc kubenswrapper[4894]: I1209 15:49:48.694132 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/horizon-6b9bcbff6b-l6n7z" podUID="6b8de55a-230d-4dbe-9f6c-aae6138d865f" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.144:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.144:8443: connect: connection refused" Dec 09 15:49:51 crc kubenswrapper[4894]: I1209 15:49:51.954960 4894 generic.go:334] "Generic (PLEG): container finished" podID="d29fb637-ba21-4cf9-ae4d-cd9720dd5af4" containerID="7a1928ef8d06e25be6e0fab76f6d003eaa7e0f5678729dc4598fa1ebcfbe1d67" exitCode=0 Dec 09 15:49:51 crc kubenswrapper[4894]: I1209 15:49:51.955003 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-s4wr5" event={"ID":"d29fb637-ba21-4cf9-ae4d-cd9720dd5af4","Type":"ContainerDied","Data":"7a1928ef8d06e25be6e0fab76f6d003eaa7e0f5678729dc4598fa1ebcfbe1d67"} Dec 09 15:49:52 crc kubenswrapper[4894]: I1209 15:49:52.967902 4894 generic.go:334] "Generic (PLEG): container finished" podID="ce9f3ab9-097d-4d0d-8082-08019bce834c" containerID="f2e5fa88e421d711a8f091021fdd47f49dfbfe3a8d6e38495324d1020a7b71ea" exitCode=0 Dec 09 15:49:52 crc kubenswrapper[4894]: I1209 15:49:52.968154 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-nbnz8" event={"ID":"ce9f3ab9-097d-4d0d-8082-08019bce834c","Type":"ContainerDied","Data":"f2e5fa88e421d711a8f091021fdd47f49dfbfe3a8d6e38495324d1020a7b71ea"} Dec 09 15:49:52 crc kubenswrapper[4894]: I1209 15:49:52.979840 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bab95882-60a4-4e02-b3b7-b2a6ae55315f","Type":"ContainerStarted","Data":"2a84f533d78019c9f713e6d77e8cd58d1ca1eee10176c65b3466d16f959bb39e"} Dec 09 15:49:52 crc kubenswrapper[4894]: I1209 15:49:52.979965 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bab95882-60a4-4e02-b3b7-b2a6ae55315f" containerName="ceilometer-central-agent" containerID="cri-o://42832b08996b48d3dee186127faf58ac23a8ebe1bcccbb9bc653bbc535ff663c" gracePeriod=30 Dec 09 15:49:52 crc kubenswrapper[4894]: I1209 15:49:52.980238 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bab95882-60a4-4e02-b3b7-b2a6ae55315f" containerName="proxy-httpd" containerID="cri-o://2a84f533d78019c9f713e6d77e8cd58d1ca1eee10176c65b3466d16f959bb39e" gracePeriod=30 Dec 09 15:49:52 crc kubenswrapper[4894]: I1209 15:49:52.980338 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bab95882-60a4-4e02-b3b7-b2a6ae55315f" containerName="sg-core" containerID="cri-o://36c3822c1e6a90d261e05cba628235eab698bf1a47839de9a35bce0fe9dc2b06" gracePeriod=30 Dec 09 15:49:52 crc kubenswrapper[4894]: I1209 15:49:52.980388 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="bab95882-60a4-4e02-b3b7-b2a6ae55315f" containerName="ceilometer-notification-agent" containerID="cri-o://ad993b83ca0bae3b9d2a6aa998c2747bd8ca7896fb6279c3e80ae74e1c99bd1f" gracePeriod=30 Dec 09 15:49:53 crc kubenswrapper[4894]: I1209 15:49:53.015489 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.307727745 podStartE2EDuration="54.01547011s" podCreationTimestamp="2025-12-09 15:48:59 +0000 UTC" firstStartedPulling="2025-12-09 15:49:00.443418015 +0000 UTC m=+1034.762628674" lastFinishedPulling="2025-12-09 15:49:52.151160359 +0000 UTC m=+1086.470371039" observedRunningTime="2025-12-09 15:49:53.015196363 +0000 UTC m=+1087.334407022" watchObservedRunningTime="2025-12-09 15:49:53.01547011 +0000 UTC m=+1087.334680779" Dec 09 15:49:53 crc kubenswrapper[4894]: I1209 15:49:53.278155 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-s4wr5" Dec 09 15:49:53 crc kubenswrapper[4894]: I1209 15:49:53.401540 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29fb637-ba21-4cf9-ae4d-cd9720dd5af4-combined-ca-bundle\") pod \"d29fb637-ba21-4cf9-ae4d-cd9720dd5af4\" (UID: \"d29fb637-ba21-4cf9-ae4d-cd9720dd5af4\") " Dec 09 15:49:53 crc kubenswrapper[4894]: I1209 15:49:53.401591 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b4rm6\" (UniqueName: \"kubernetes.io/projected/d29fb637-ba21-4cf9-ae4d-cd9720dd5af4-kube-api-access-b4rm6\") pod \"d29fb637-ba21-4cf9-ae4d-cd9720dd5af4\" (UID: \"d29fb637-ba21-4cf9-ae4d-cd9720dd5af4\") " Dec 09 15:49:53 crc kubenswrapper[4894]: I1209 15:49:53.401711 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d29fb637-ba21-4cf9-ae4d-cd9720dd5af4-db-sync-config-data\") pod \"d29fb637-ba21-4cf9-ae4d-cd9720dd5af4\" (UID: \"d29fb637-ba21-4cf9-ae4d-cd9720dd5af4\") " Dec 09 15:49:53 crc kubenswrapper[4894]: I1209 15:49:53.407211 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d29fb637-ba21-4cf9-ae4d-cd9720dd5af4-kube-api-access-b4rm6" (OuterVolumeSpecName: "kube-api-access-b4rm6") pod "d29fb637-ba21-4cf9-ae4d-cd9720dd5af4" (UID: "d29fb637-ba21-4cf9-ae4d-cd9720dd5af4"). InnerVolumeSpecName "kube-api-access-b4rm6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:49:53 crc kubenswrapper[4894]: I1209 15:49:53.407410 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d29fb637-ba21-4cf9-ae4d-cd9720dd5af4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "d29fb637-ba21-4cf9-ae4d-cd9720dd5af4" (UID: "d29fb637-ba21-4cf9-ae4d-cd9720dd5af4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:53 crc kubenswrapper[4894]: I1209 15:49:53.428914 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d29fb637-ba21-4cf9-ae4d-cd9720dd5af4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d29fb637-ba21-4cf9-ae4d-cd9720dd5af4" (UID: "d29fb637-ba21-4cf9-ae4d-cd9720dd5af4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:53 crc kubenswrapper[4894]: I1209 15:49:53.503669 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d29fb637-ba21-4cf9-ae4d-cd9720dd5af4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:53 crc kubenswrapper[4894]: I1209 15:49:53.504665 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b4rm6\" (UniqueName: \"kubernetes.io/projected/d29fb637-ba21-4cf9-ae4d-cd9720dd5af4-kube-api-access-b4rm6\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:53 crc kubenswrapper[4894]: I1209 15:49:53.504717 4894 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/d29fb637-ba21-4cf9-ae4d-cd9720dd5af4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:53 crc kubenswrapper[4894]: I1209 15:49:53.997875 4894 generic.go:334] "Generic (PLEG): container finished" podID="bab95882-60a4-4e02-b3b7-b2a6ae55315f" containerID="2a84f533d78019c9f713e6d77e8cd58d1ca1eee10176c65b3466d16f959bb39e" exitCode=0 Dec 09 15:49:53 crc kubenswrapper[4894]: I1209 15:49:53.998175 4894 generic.go:334] "Generic (PLEG): container finished" podID="bab95882-60a4-4e02-b3b7-b2a6ae55315f" containerID="36c3822c1e6a90d261e05cba628235eab698bf1a47839de9a35bce0fe9dc2b06" exitCode=2 Dec 09 15:49:53 crc kubenswrapper[4894]: I1209 15:49:53.998190 4894 generic.go:334] "Generic (PLEG): container finished" podID="bab95882-60a4-4e02-b3b7-b2a6ae55315f" containerID="42832b08996b48d3dee186127faf58ac23a8ebe1bcccbb9bc653bbc535ff663c" exitCode=0 Dec 09 15:49:53 crc kubenswrapper[4894]: I1209 15:49:53.998111 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bab95882-60a4-4e02-b3b7-b2a6ae55315f","Type":"ContainerDied","Data":"2a84f533d78019c9f713e6d77e8cd58d1ca1eee10176c65b3466d16f959bb39e"} Dec 09 15:49:53 crc kubenswrapper[4894]: I1209 15:49:53.998275 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bab95882-60a4-4e02-b3b7-b2a6ae55315f","Type":"ContainerDied","Data":"36c3822c1e6a90d261e05cba628235eab698bf1a47839de9a35bce0fe9dc2b06"} Dec 09 15:49:53 crc kubenswrapper[4894]: I1209 15:49:53.998289 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bab95882-60a4-4e02-b3b7-b2a6ae55315f","Type":"ContainerDied","Data":"42832b08996b48d3dee186127faf58ac23a8ebe1bcccbb9bc653bbc535ff663c"} Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.000651 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-s4wr5" event={"ID":"d29fb637-ba21-4cf9-ae4d-cd9720dd5af4","Type":"ContainerDied","Data":"f532e366948e0e6b1b702f291b0c0e8660fce048f415d2b7215f7a995067200c"} Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.000685 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-s4wr5" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.000690 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f532e366948e0e6b1b702f291b0c0e8660fce048f415d2b7215f7a995067200c" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.167772 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-5776b569cc-b4c82"] Dec 09 15:49:54 crc kubenswrapper[4894]: E1209 15:49:54.168111 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d29fb637-ba21-4cf9-ae4d-cd9720dd5af4" containerName="barbican-db-sync" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.168126 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d29fb637-ba21-4cf9-ae4d-cd9720dd5af4" containerName="barbican-db-sync" Dec 09 15:49:54 crc kubenswrapper[4894]: E1209 15:49:54.168143 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f20c656d-4752-41ce-a1aa-aa6980430e90" containerName="dnsmasq-dns" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.168149 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f20c656d-4752-41ce-a1aa-aa6980430e90" containerName="dnsmasq-dns" Dec 09 15:49:54 crc kubenswrapper[4894]: E1209 15:49:54.168162 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f20c656d-4752-41ce-a1aa-aa6980430e90" containerName="init" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.168168 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f20c656d-4752-41ce-a1aa-aa6980430e90" containerName="init" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.168335 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="d29fb637-ba21-4cf9-ae4d-cd9720dd5af4" containerName="barbican-db-sync" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.168361 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f20c656d-4752-41ce-a1aa-aa6980430e90" containerName="dnsmasq-dns" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.169216 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5776b569cc-b4c82" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.172520 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.173272 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.173620 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-x8k6m" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.180339 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-67f4558b8d-v6cpx"] Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.182368 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-67f4558b8d-v6cpx" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.184991 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.195417 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5776b569cc-b4c82"] Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.239857 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-67f4558b8d-v6cpx"] Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.305331 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-75479fdbd9-z4s7v"] Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.309889 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75479fdbd9-z4s7v" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.318877 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-75479fdbd9-z4s7v"] Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.319649 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b470d89f-2e08-4222-adc9-5dab929008ea-logs\") pod \"barbican-keystone-listener-67f4558b8d-v6cpx\" (UID: \"b470d89f-2e08-4222-adc9-5dab929008ea\") " pod="openstack/barbican-keystone-listener-67f4558b8d-v6cpx" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.319940 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b470d89f-2e08-4222-adc9-5dab929008ea-combined-ca-bundle\") pod \"barbican-keystone-listener-67f4558b8d-v6cpx\" (UID: \"b470d89f-2e08-4222-adc9-5dab929008ea\") " pod="openstack/barbican-keystone-listener-67f4558b8d-v6cpx" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.320001 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a921fbc-ca69-439e-b357-eb15cc4de86e-config-data\") pod \"barbican-worker-5776b569cc-b4c82\" (UID: \"7a921fbc-ca69-439e-b357-eb15cc4de86e\") " pod="openstack/barbican-worker-5776b569cc-b4c82" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.320036 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vtck9\" (UniqueName: \"kubernetes.io/projected/7a921fbc-ca69-439e-b357-eb15cc4de86e-kube-api-access-vtck9\") pod \"barbican-worker-5776b569cc-b4c82\" (UID: \"7a921fbc-ca69-439e-b357-eb15cc4de86e\") " pod="openstack/barbican-worker-5776b569cc-b4c82" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.320069 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7a921fbc-ca69-439e-b357-eb15cc4de86e-config-data-custom\") pod \"barbican-worker-5776b569cc-b4c82\" (UID: \"7a921fbc-ca69-439e-b357-eb15cc4de86e\") " pod="openstack/barbican-worker-5776b569cc-b4c82" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.320094 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b470d89f-2e08-4222-adc9-5dab929008ea-config-data\") pod \"barbican-keystone-listener-67f4558b8d-v6cpx\" (UID: \"b470d89f-2e08-4222-adc9-5dab929008ea\") " pod="openstack/barbican-keystone-listener-67f4558b8d-v6cpx" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.320145 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a921fbc-ca69-439e-b357-eb15cc4de86e-combined-ca-bundle\") pod \"barbican-worker-5776b569cc-b4c82\" (UID: \"7a921fbc-ca69-439e-b357-eb15cc4de86e\") " pod="openstack/barbican-worker-5776b569cc-b4c82" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.320181 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a921fbc-ca69-439e-b357-eb15cc4de86e-logs\") pod \"barbican-worker-5776b569cc-b4c82\" (UID: \"7a921fbc-ca69-439e-b357-eb15cc4de86e\") " pod="openstack/barbican-worker-5776b569cc-b4c82" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.320213 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b470d89f-2e08-4222-adc9-5dab929008ea-config-data-custom\") pod \"barbican-keystone-listener-67f4558b8d-v6cpx\" (UID: \"b470d89f-2e08-4222-adc9-5dab929008ea\") " pod="openstack/barbican-keystone-listener-67f4558b8d-v6cpx" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.320237 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jhkd8\" (UniqueName: \"kubernetes.io/projected/b470d89f-2e08-4222-adc9-5dab929008ea-kube-api-access-jhkd8\") pod \"barbican-keystone-listener-67f4558b8d-v6cpx\" (UID: \"b470d89f-2e08-4222-adc9-5dab929008ea\") " pod="openstack/barbican-keystone-listener-67f4558b8d-v6cpx" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.414520 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-569f9c8ccb-pkfj6"] Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.415799 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.418403 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.421829 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a921fbc-ca69-439e-b357-eb15cc4de86e-logs\") pod \"barbican-worker-5776b569cc-b4c82\" (UID: \"7a921fbc-ca69-439e-b357-eb15cc4de86e\") " pod="openstack/barbican-worker-5776b569cc-b4c82" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.421867 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b470d89f-2e08-4222-adc9-5dab929008ea-config-data-custom\") pod \"barbican-keystone-listener-67f4558b8d-v6cpx\" (UID: \"b470d89f-2e08-4222-adc9-5dab929008ea\") " pod="openstack/barbican-keystone-listener-67f4558b8d-v6cpx" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.421891 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jhkd8\" (UniqueName: \"kubernetes.io/projected/b470d89f-2e08-4222-adc9-5dab929008ea-kube-api-access-jhkd8\") pod \"barbican-keystone-listener-67f4558b8d-v6cpx\" (UID: \"b470d89f-2e08-4222-adc9-5dab929008ea\") " pod="openstack/barbican-keystone-listener-67f4558b8d-v6cpx" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.421921 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mbm4g\" (UniqueName: \"kubernetes.io/projected/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-kube-api-access-mbm4g\") pod \"dnsmasq-dns-75479fdbd9-z4s7v\" (UID: \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\") " pod="openstack/dnsmasq-dns-75479fdbd9-z4s7v" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.421943 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-ovsdbserver-nb\") pod \"dnsmasq-dns-75479fdbd9-z4s7v\" (UID: \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\") " pod="openstack/dnsmasq-dns-75479fdbd9-z4s7v" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.421966 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b470d89f-2e08-4222-adc9-5dab929008ea-logs\") pod \"barbican-keystone-listener-67f4558b8d-v6cpx\" (UID: \"b470d89f-2e08-4222-adc9-5dab929008ea\") " pod="openstack/barbican-keystone-listener-67f4558b8d-v6cpx" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.422003 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-config-data\") pod \"barbican-api-569f9c8ccb-pkfj6\" (UID: \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\") " pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.422024 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-config-data-custom\") pod \"barbican-api-569f9c8ccb-pkfj6\" (UID: \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\") " pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.422047 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b470d89f-2e08-4222-adc9-5dab929008ea-combined-ca-bundle\") pod \"barbican-keystone-listener-67f4558b8d-v6cpx\" (UID: \"b470d89f-2e08-4222-adc9-5dab929008ea\") " pod="openstack/barbican-keystone-listener-67f4558b8d-v6cpx" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.422068 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-ovsdbserver-sb\") pod \"dnsmasq-dns-75479fdbd9-z4s7v\" (UID: \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\") " pod="openstack/dnsmasq-dns-75479fdbd9-z4s7v" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.422087 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sr87v\" (UniqueName: \"kubernetes.io/projected/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-kube-api-access-sr87v\") pod \"barbican-api-569f9c8ccb-pkfj6\" (UID: \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\") " pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.422111 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a921fbc-ca69-439e-b357-eb15cc4de86e-config-data\") pod \"barbican-worker-5776b569cc-b4c82\" (UID: \"7a921fbc-ca69-439e-b357-eb15cc4de86e\") " pod="openstack/barbican-worker-5776b569cc-b4c82" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.422127 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-logs\") pod \"barbican-api-569f9c8ccb-pkfj6\" (UID: \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\") " pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.422145 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vtck9\" (UniqueName: \"kubernetes.io/projected/7a921fbc-ca69-439e-b357-eb15cc4de86e-kube-api-access-vtck9\") pod \"barbican-worker-5776b569cc-b4c82\" (UID: \"7a921fbc-ca69-439e-b357-eb15cc4de86e\") " pod="openstack/barbican-worker-5776b569cc-b4c82" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.422160 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-config\") pod \"dnsmasq-dns-75479fdbd9-z4s7v\" (UID: \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\") " pod="openstack/dnsmasq-dns-75479fdbd9-z4s7v" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.422181 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7a921fbc-ca69-439e-b357-eb15cc4de86e-config-data-custom\") pod \"barbican-worker-5776b569cc-b4c82\" (UID: \"7a921fbc-ca69-439e-b357-eb15cc4de86e\") " pod="openstack/barbican-worker-5776b569cc-b4c82" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.422200 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b470d89f-2e08-4222-adc9-5dab929008ea-config-data\") pod \"barbican-keystone-listener-67f4558b8d-v6cpx\" (UID: \"b470d89f-2e08-4222-adc9-5dab929008ea\") " pod="openstack/barbican-keystone-listener-67f4558b8d-v6cpx" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.422218 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-combined-ca-bundle\") pod \"barbican-api-569f9c8ccb-pkfj6\" (UID: \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\") " pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.422253 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a921fbc-ca69-439e-b357-eb15cc4de86e-combined-ca-bundle\") pod \"barbican-worker-5776b569cc-b4c82\" (UID: \"7a921fbc-ca69-439e-b357-eb15cc4de86e\") " pod="openstack/barbican-worker-5776b569cc-b4c82" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.422271 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-dns-svc\") pod \"dnsmasq-dns-75479fdbd9-z4s7v\" (UID: \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\") " pod="openstack/dnsmasq-dns-75479fdbd9-z4s7v" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.422737 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7a921fbc-ca69-439e-b357-eb15cc4de86e-logs\") pod \"barbican-worker-5776b569cc-b4c82\" (UID: \"7a921fbc-ca69-439e-b357-eb15cc4de86e\") " pod="openstack/barbican-worker-5776b569cc-b4c82" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.427278 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b470d89f-2e08-4222-adc9-5dab929008ea-logs\") pod \"barbican-keystone-listener-67f4558b8d-v6cpx\" (UID: \"b470d89f-2e08-4222-adc9-5dab929008ea\") " pod="openstack/barbican-keystone-listener-67f4558b8d-v6cpx" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.429489 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/7a921fbc-ca69-439e-b357-eb15cc4de86e-config-data-custom\") pod \"barbican-worker-5776b569cc-b4c82\" (UID: \"7a921fbc-ca69-439e-b357-eb15cc4de86e\") " pod="openstack/barbican-worker-5776b569cc-b4c82" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.432202 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7a921fbc-ca69-439e-b357-eb15cc4de86e-combined-ca-bundle\") pod \"barbican-worker-5776b569cc-b4c82\" (UID: \"7a921fbc-ca69-439e-b357-eb15cc4de86e\") " pod="openstack/barbican-worker-5776b569cc-b4c82" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.434431 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b470d89f-2e08-4222-adc9-5dab929008ea-combined-ca-bundle\") pod \"barbican-keystone-listener-67f4558b8d-v6cpx\" (UID: \"b470d89f-2e08-4222-adc9-5dab929008ea\") " pod="openstack/barbican-keystone-listener-67f4558b8d-v6cpx" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.438659 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b470d89f-2e08-4222-adc9-5dab929008ea-config-data\") pod \"barbican-keystone-listener-67f4558b8d-v6cpx\" (UID: \"b470d89f-2e08-4222-adc9-5dab929008ea\") " pod="openstack/barbican-keystone-listener-67f4558b8d-v6cpx" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.442513 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/b470d89f-2e08-4222-adc9-5dab929008ea-config-data-custom\") pod \"barbican-keystone-listener-67f4558b8d-v6cpx\" (UID: \"b470d89f-2e08-4222-adc9-5dab929008ea\") " pod="openstack/barbican-keystone-listener-67f4558b8d-v6cpx" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.445737 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vtck9\" (UniqueName: \"kubernetes.io/projected/7a921fbc-ca69-439e-b357-eb15cc4de86e-kube-api-access-vtck9\") pod \"barbican-worker-5776b569cc-b4c82\" (UID: \"7a921fbc-ca69-439e-b357-eb15cc4de86e\") " pod="openstack/barbican-worker-5776b569cc-b4c82" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.452875 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7a921fbc-ca69-439e-b357-eb15cc4de86e-config-data\") pod \"barbican-worker-5776b569cc-b4c82\" (UID: \"7a921fbc-ca69-439e-b357-eb15cc4de86e\") " pod="openstack/barbican-worker-5776b569cc-b4c82" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.453410 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jhkd8\" (UniqueName: \"kubernetes.io/projected/b470d89f-2e08-4222-adc9-5dab929008ea-kube-api-access-jhkd8\") pod \"barbican-keystone-listener-67f4558b8d-v6cpx\" (UID: \"b470d89f-2e08-4222-adc9-5dab929008ea\") " pod="openstack/barbican-keystone-listener-67f4558b8d-v6cpx" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.455956 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-569f9c8ccb-pkfj6"] Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.493795 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-5776b569cc-b4c82" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.509062 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-67f4558b8d-v6cpx" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.525183 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-config\") pod \"dnsmasq-dns-75479fdbd9-z4s7v\" (UID: \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\") " pod="openstack/dnsmasq-dns-75479fdbd9-z4s7v" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.525237 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-combined-ca-bundle\") pod \"barbican-api-569f9c8ccb-pkfj6\" (UID: \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\") " pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.525266 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-dns-svc\") pod \"dnsmasq-dns-75479fdbd9-z4s7v\" (UID: \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\") " pod="openstack/dnsmasq-dns-75479fdbd9-z4s7v" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.525540 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mbm4g\" (UniqueName: \"kubernetes.io/projected/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-kube-api-access-mbm4g\") pod \"dnsmasq-dns-75479fdbd9-z4s7v\" (UID: \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\") " pod="openstack/dnsmasq-dns-75479fdbd9-z4s7v" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.525584 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-ovsdbserver-nb\") pod \"dnsmasq-dns-75479fdbd9-z4s7v\" (UID: \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\") " pod="openstack/dnsmasq-dns-75479fdbd9-z4s7v" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.526871 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-ovsdbserver-nb\") pod \"dnsmasq-dns-75479fdbd9-z4s7v\" (UID: \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\") " pod="openstack/dnsmasq-dns-75479fdbd9-z4s7v" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.526990 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-config\") pod \"dnsmasq-dns-75479fdbd9-z4s7v\" (UID: \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\") " pod="openstack/dnsmasq-dns-75479fdbd9-z4s7v" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.527027 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-config-data\") pod \"barbican-api-569f9c8ccb-pkfj6\" (UID: \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\") " pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.527521 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-dns-svc\") pod \"dnsmasq-dns-75479fdbd9-z4s7v\" (UID: \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\") " pod="openstack/dnsmasq-dns-75479fdbd9-z4s7v" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.527675 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-config-data-custom\") pod \"barbican-api-569f9c8ccb-pkfj6\" (UID: \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\") " pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.527740 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-ovsdbserver-sb\") pod \"dnsmasq-dns-75479fdbd9-z4s7v\" (UID: \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\") " pod="openstack/dnsmasq-dns-75479fdbd9-z4s7v" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.527763 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sr87v\" (UniqueName: \"kubernetes.io/projected/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-kube-api-access-sr87v\") pod \"barbican-api-569f9c8ccb-pkfj6\" (UID: \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\") " pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.529355 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-logs\") pod \"barbican-api-569f9c8ccb-pkfj6\" (UID: \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\") " pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.529808 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-logs\") pod \"barbican-api-569f9c8ccb-pkfj6\" (UID: \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\") " pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.530999 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-nbnz8" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.533365 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-ovsdbserver-sb\") pod \"dnsmasq-dns-75479fdbd9-z4s7v\" (UID: \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\") " pod="openstack/dnsmasq-dns-75479fdbd9-z4s7v" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.541029 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-config-data-custom\") pod \"barbican-api-569f9c8ccb-pkfj6\" (UID: \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\") " pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.542062 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-config-data\") pod \"barbican-api-569f9c8ccb-pkfj6\" (UID: \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\") " pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.542077 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-combined-ca-bundle\") pod \"barbican-api-569f9c8ccb-pkfj6\" (UID: \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\") " pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.557067 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sr87v\" (UniqueName: \"kubernetes.io/projected/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-kube-api-access-sr87v\") pod \"barbican-api-569f9c8ccb-pkfj6\" (UID: \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\") " pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.557687 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mbm4g\" (UniqueName: \"kubernetes.io/projected/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-kube-api-access-mbm4g\") pod \"dnsmasq-dns-75479fdbd9-z4s7v\" (UID: \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\") " pod="openstack/dnsmasq-dns-75479fdbd9-z4s7v" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.614320 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.630881 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-combined-ca-bundle\") pod \"ce9f3ab9-097d-4d0d-8082-08019bce834c\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.630961 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ce9f3ab9-097d-4d0d-8082-08019bce834c-etc-machine-id\") pod \"ce9f3ab9-097d-4d0d-8082-08019bce834c\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.630995 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-db-sync-config-data\") pod \"ce9f3ab9-097d-4d0d-8082-08019bce834c\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.631265 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-config-data\") pod \"ce9f3ab9-097d-4d0d-8082-08019bce834c\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.631333 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-scripts\") pod \"ce9f3ab9-097d-4d0d-8082-08019bce834c\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.631406 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wwj2\" (UniqueName: \"kubernetes.io/projected/ce9f3ab9-097d-4d0d-8082-08019bce834c-kube-api-access-6wwj2\") pod \"ce9f3ab9-097d-4d0d-8082-08019bce834c\" (UID: \"ce9f3ab9-097d-4d0d-8082-08019bce834c\") " Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.632774 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ce9f3ab9-097d-4d0d-8082-08019bce834c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "ce9f3ab9-097d-4d0d-8082-08019bce834c" (UID: "ce9f3ab9-097d-4d0d-8082-08019bce834c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.633944 4894 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/ce9f3ab9-097d-4d0d-8082-08019bce834c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.639529 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-scripts" (OuterVolumeSpecName: "scripts") pod "ce9f3ab9-097d-4d0d-8082-08019bce834c" (UID: "ce9f3ab9-097d-4d0d-8082-08019bce834c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.644360 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "ce9f3ab9-097d-4d0d-8082-08019bce834c" (UID: "ce9f3ab9-097d-4d0d-8082-08019bce834c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.650408 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75479fdbd9-z4s7v" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.650451 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ce9f3ab9-097d-4d0d-8082-08019bce834c-kube-api-access-6wwj2" (OuterVolumeSpecName: "kube-api-access-6wwj2") pod "ce9f3ab9-097d-4d0d-8082-08019bce834c" (UID: "ce9f3ab9-097d-4d0d-8082-08019bce834c"). InnerVolumeSpecName "kube-api-access-6wwj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.697608 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ce9f3ab9-097d-4d0d-8082-08019bce834c" (UID: "ce9f3ab9-097d-4d0d-8082-08019bce834c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.712838 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-config-data" (OuterVolumeSpecName: "config-data") pod "ce9f3ab9-097d-4d0d-8082-08019bce834c" (UID: "ce9f3ab9-097d-4d0d-8082-08019bce834c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.738535 4894 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.738572 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.738587 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.738614 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wwj2\" (UniqueName: \"kubernetes.io/projected/ce9f3ab9-097d-4d0d-8082-08019bce834c-kube-api-access-6wwj2\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:54 crc kubenswrapper[4894]: I1209 15:49:54.738630 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ce9f3ab9-097d-4d0d-8082-08019bce834c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.010560 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-nbnz8" event={"ID":"ce9f3ab9-097d-4d0d-8082-08019bce834c","Type":"ContainerDied","Data":"16473a4acae940ea360630bcbed55206fc73095e0003fd944f4461f7506d5cff"} Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.010603 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16473a4acae940ea360630bcbed55206fc73095e0003fd944f4461f7506d5cff" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.010605 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-nbnz8" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.020626 4894 generic.go:334] "Generic (PLEG): container finished" podID="bab95882-60a4-4e02-b3b7-b2a6ae55315f" containerID="ad993b83ca0bae3b9d2a6aa998c2747bd8ca7896fb6279c3e80ae74e1c99bd1f" exitCode=0 Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.020684 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bab95882-60a4-4e02-b3b7-b2a6ae55315f","Type":"ContainerDied","Data":"ad993b83ca0bae3b9d2a6aa998c2747bd8ca7896fb6279c3e80ae74e1c99bd1f"} Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.051713 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-5776b569cc-b4c82"] Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.302879 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 15:49:55 crc kubenswrapper[4894]: E1209 15:49:55.303464 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ce9f3ab9-097d-4d0d-8082-08019bce834c" containerName="cinder-db-sync" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.303476 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ce9f3ab9-097d-4d0d-8082-08019bce834c" containerName="cinder-db-sync" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.303754 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="ce9f3ab9-097d-4d0d-8082-08019bce834c" containerName="cinder-db-sync" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.304632 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.307540 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.308022 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.308138 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.308904 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-ktrhh" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.326441 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.351051 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " pod="openstack/cinder-scheduler-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.351103 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " pod="openstack/cinder-scheduler-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.351127 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8c380b2-49de-4630-80e9-15c73cd99045-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " pod="openstack/cinder-scheduler-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.351150 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-scripts\") pod \"cinder-scheduler-0\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " pod="openstack/cinder-scheduler-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.351223 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqp46\" (UniqueName: \"kubernetes.io/projected/f8c380b2-49de-4630-80e9-15c73cd99045-kube-api-access-qqp46\") pod \"cinder-scheduler-0\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " pod="openstack/cinder-scheduler-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.351263 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-config-data\") pod \"cinder-scheduler-0\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " pod="openstack/cinder-scheduler-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.353187 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75479fdbd9-z4s7v"] Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.428296 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-68c677b759-g4lpm"] Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.430032 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68c677b759-g4lpm" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.437202 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68c677b759-g4lpm"] Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.465202 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " pod="openstack/cinder-scheduler-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.465248 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-ovsdbserver-sb\") pod \"dnsmasq-dns-68c677b759-g4lpm\" (UID: \"d3f62001-a311-45d1-ab65-15cd3de84d4e\") " pod="openstack/dnsmasq-dns-68c677b759-g4lpm" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.465273 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8c380b2-49de-4630-80e9-15c73cd99045-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " pod="openstack/cinder-scheduler-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.465298 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-scripts\") pod \"cinder-scheduler-0\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " pod="openstack/cinder-scheduler-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.465321 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-dns-svc\") pod \"dnsmasq-dns-68c677b759-g4lpm\" (UID: \"d3f62001-a311-45d1-ab65-15cd3de84d4e\") " pod="openstack/dnsmasq-dns-68c677b759-g4lpm" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.465373 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqp46\" (UniqueName: \"kubernetes.io/projected/f8c380b2-49de-4630-80e9-15c73cd99045-kube-api-access-qqp46\") pod \"cinder-scheduler-0\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " pod="openstack/cinder-scheduler-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.465396 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-config\") pod \"dnsmasq-dns-68c677b759-g4lpm\" (UID: \"d3f62001-a311-45d1-ab65-15cd3de84d4e\") " pod="openstack/dnsmasq-dns-68c677b759-g4lpm" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.465424 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-config-data\") pod \"cinder-scheduler-0\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " pod="openstack/cinder-scheduler-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.465479 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-ovsdbserver-nb\") pod \"dnsmasq-dns-68c677b759-g4lpm\" (UID: \"d3f62001-a311-45d1-ab65-15cd3de84d4e\") " pod="openstack/dnsmasq-dns-68c677b759-g4lpm" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.465496 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " pod="openstack/cinder-scheduler-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.465517 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h6b7\" (UniqueName: \"kubernetes.io/projected/d3f62001-a311-45d1-ab65-15cd3de84d4e-kube-api-access-8h6b7\") pod \"dnsmasq-dns-68c677b759-g4lpm\" (UID: \"d3f62001-a311-45d1-ab65-15cd3de84d4e\") " pod="openstack/dnsmasq-dns-68c677b759-g4lpm" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.465863 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8c380b2-49de-4630-80e9-15c73cd99045-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " pod="openstack/cinder-scheduler-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.479616 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " pod="openstack/cinder-scheduler-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.485822 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " pod="openstack/cinder-scheduler-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.490363 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-scripts\") pod \"cinder-scheduler-0\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " pod="openstack/cinder-scheduler-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.495299 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqp46\" (UniqueName: \"kubernetes.io/projected/f8c380b2-49de-4630-80e9-15c73cd99045-kube-api-access-qqp46\") pod \"cinder-scheduler-0\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " pod="openstack/cinder-scheduler-0" Dec 09 15:49:55 crc kubenswrapper[4894]: W1209 15:49:55.495877 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5ecec7eb_d196_471a_9e5c_4ae0ce5ee992.slice/crio-4d4aed6ba4c61e4e05b19fe908160be8df98cfde903dbb0f43021286e99abe42 WatchSource:0}: Error finding container 4d4aed6ba4c61e4e05b19fe908160be8df98cfde903dbb0f43021286e99abe42: Status 404 returned error can't find the container with id 4d4aed6ba4c61e4e05b19fe908160be8df98cfde903dbb0f43021286e99abe42 Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.505586 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-config-data\") pod \"cinder-scheduler-0\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " pod="openstack/cinder-scheduler-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.509339 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-569f9c8ccb-pkfj6"] Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.526743 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-67f4558b8d-v6cpx"] Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.556474 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.558297 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.567318 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-config\") pod \"dnsmasq-dns-68c677b759-g4lpm\" (UID: \"d3f62001-a311-45d1-ab65-15cd3de84d4e\") " pod="openstack/dnsmasq-dns-68c677b759-g4lpm" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.567528 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-ovsdbserver-nb\") pod \"dnsmasq-dns-68c677b759-g4lpm\" (UID: \"d3f62001-a311-45d1-ab65-15cd3de84d4e\") " pod="openstack/dnsmasq-dns-68c677b759-g4lpm" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.567602 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h6b7\" (UniqueName: \"kubernetes.io/projected/d3f62001-a311-45d1-ab65-15cd3de84d4e-kube-api-access-8h6b7\") pod \"dnsmasq-dns-68c677b759-g4lpm\" (UID: \"d3f62001-a311-45d1-ab65-15cd3de84d4e\") " pod="openstack/dnsmasq-dns-68c677b759-g4lpm" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.567740 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-ovsdbserver-sb\") pod \"dnsmasq-dns-68c677b759-g4lpm\" (UID: \"d3f62001-a311-45d1-ab65-15cd3de84d4e\") " pod="openstack/dnsmasq-dns-68c677b759-g4lpm" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.567796 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-dns-svc\") pod \"dnsmasq-dns-68c677b759-g4lpm\" (UID: \"d3f62001-a311-45d1-ab65-15cd3de84d4e\") " pod="openstack/dnsmasq-dns-68c677b759-g4lpm" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.567980 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.568961 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-dns-svc\") pod \"dnsmasq-dns-68c677b759-g4lpm\" (UID: \"d3f62001-a311-45d1-ab65-15cd3de84d4e\") " pod="openstack/dnsmasq-dns-68c677b759-g4lpm" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.569186 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-ovsdbserver-nb\") pod \"dnsmasq-dns-68c677b759-g4lpm\" (UID: \"d3f62001-a311-45d1-ab65-15cd3de84d4e\") " pod="openstack/dnsmasq-dns-68c677b759-g4lpm" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.569759 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-config\") pod \"dnsmasq-dns-68c677b759-g4lpm\" (UID: \"d3f62001-a311-45d1-ab65-15cd3de84d4e\") " pod="openstack/dnsmasq-dns-68c677b759-g4lpm" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.570381 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-ovsdbserver-sb\") pod \"dnsmasq-dns-68c677b759-g4lpm\" (UID: \"d3f62001-a311-45d1-ab65-15cd3de84d4e\") " pod="openstack/dnsmasq-dns-68c677b759-g4lpm" Dec 09 15:49:55 crc kubenswrapper[4894]: W1209 15:49:55.571939 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfee9920a_8d2a_4bf3_88dc_5d745c0dee82.slice/crio-ae4d186a57c63f678215acc088926fa2856567f4b8f325c85444e18740e96676 WatchSource:0}: Error finding container ae4d186a57c63f678215acc088926fa2856567f4b8f325c85444e18740e96676: Status 404 returned error can't find the container with id ae4d186a57c63f678215acc088926fa2856567f4b8f325c85444e18740e96676 Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.585305 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75479fdbd9-z4s7v"] Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.595842 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h6b7\" (UniqueName: \"kubernetes.io/projected/d3f62001-a311-45d1-ab65-15cd3de84d4e-kube-api-access-8h6b7\") pod \"dnsmasq-dns-68c677b759-g4lpm\" (UID: \"d3f62001-a311-45d1-ab65-15cd3de84d4e\") " pod="openstack/dnsmasq-dns-68c677b759-g4lpm" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.602769 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.644154 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.669160 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-scripts\") pod \"cinder-api-0\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " pod="openstack/cinder-api-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.669228 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6n4vb\" (UniqueName: \"kubernetes.io/projected/e04dd18f-5e7f-464b-9f65-d4233a0753a6-kube-api-access-6n4vb\") pod \"cinder-api-0\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " pod="openstack/cinder-api-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.669254 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e04dd18f-5e7f-464b-9f65-d4233a0753a6-logs\") pod \"cinder-api-0\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " pod="openstack/cinder-api-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.669279 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e04dd18f-5e7f-464b-9f65-d4233a0753a6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " pod="openstack/cinder-api-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.669369 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " pod="openstack/cinder-api-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.669448 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-config-data-custom\") pod \"cinder-api-0\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " pod="openstack/cinder-api-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.669476 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-config-data\") pod \"cinder-api-0\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " pod="openstack/cinder-api-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.680773 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68c677b759-g4lpm" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.704891 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.775058 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-sg-core-conf-yaml\") pod \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.775126 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-config-data\") pod \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.775151 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bab95882-60a4-4e02-b3b7-b2a6ae55315f-log-httpd\") pod \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.775205 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bab95882-60a4-4e02-b3b7-b2a6ae55315f-run-httpd\") pod \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.775249 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8nqvg\" (UniqueName: \"kubernetes.io/projected/bab95882-60a4-4e02-b3b7-b2a6ae55315f-kube-api-access-8nqvg\") pod \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.775311 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-combined-ca-bundle\") pod \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.775381 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-scripts\") pod \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\" (UID: \"bab95882-60a4-4e02-b3b7-b2a6ae55315f\") " Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.775598 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " pod="openstack/cinder-api-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.775711 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-config-data\") pod \"cinder-api-0\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " pod="openstack/cinder-api-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.775731 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-config-data-custom\") pod \"cinder-api-0\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " pod="openstack/cinder-api-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.775811 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-scripts\") pod \"cinder-api-0\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " pod="openstack/cinder-api-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.775840 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6n4vb\" (UniqueName: \"kubernetes.io/projected/e04dd18f-5e7f-464b-9f65-d4233a0753a6-kube-api-access-6n4vb\") pod \"cinder-api-0\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " pod="openstack/cinder-api-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.775858 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e04dd18f-5e7f-464b-9f65-d4233a0753a6-logs\") pod \"cinder-api-0\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " pod="openstack/cinder-api-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.775873 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e04dd18f-5e7f-464b-9f65-d4233a0753a6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " pod="openstack/cinder-api-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.777163 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e04dd18f-5e7f-464b-9f65-d4233a0753a6-etc-machine-id\") pod \"cinder-api-0\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " pod="openstack/cinder-api-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.777912 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e04dd18f-5e7f-464b-9f65-d4233a0753a6-logs\") pod \"cinder-api-0\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " pod="openstack/cinder-api-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.777952 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bab95882-60a4-4e02-b3b7-b2a6ae55315f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "bab95882-60a4-4e02-b3b7-b2a6ae55315f" (UID: "bab95882-60a4-4e02-b3b7-b2a6ae55315f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.778164 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bab95882-60a4-4e02-b3b7-b2a6ae55315f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "bab95882-60a4-4e02-b3b7-b2a6ae55315f" (UID: "bab95882-60a4-4e02-b3b7-b2a6ae55315f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.781661 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " pod="openstack/cinder-api-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.784322 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-scripts\") pod \"cinder-api-0\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " pod="openstack/cinder-api-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.791041 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-scripts" (OuterVolumeSpecName: "scripts") pod "bab95882-60a4-4e02-b3b7-b2a6ae55315f" (UID: "bab95882-60a4-4e02-b3b7-b2a6ae55315f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.794360 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-config-data-custom\") pod \"cinder-api-0\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " pod="openstack/cinder-api-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.794555 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-config-data\") pod \"cinder-api-0\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " pod="openstack/cinder-api-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.796963 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bab95882-60a4-4e02-b3b7-b2a6ae55315f-kube-api-access-8nqvg" (OuterVolumeSpecName: "kube-api-access-8nqvg") pod "bab95882-60a4-4e02-b3b7-b2a6ae55315f" (UID: "bab95882-60a4-4e02-b3b7-b2a6ae55315f"). InnerVolumeSpecName "kube-api-access-8nqvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.803171 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6n4vb\" (UniqueName: \"kubernetes.io/projected/e04dd18f-5e7f-464b-9f65-d4233a0753a6-kube-api-access-6n4vb\") pod \"cinder-api-0\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " pod="openstack/cinder-api-0" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.842831 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "bab95882-60a4-4e02-b3b7-b2a6ae55315f" (UID: "bab95882-60a4-4e02-b3b7-b2a6ae55315f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.878319 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bab95882-60a4-4e02-b3b7-b2a6ae55315f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.878342 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/bab95882-60a4-4e02-b3b7-b2a6ae55315f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.878351 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8nqvg\" (UniqueName: \"kubernetes.io/projected/bab95882-60a4-4e02-b3b7-b2a6ae55315f-kube-api-access-8nqvg\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.878384 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.878393 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.917070 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "bab95882-60a4-4e02-b3b7-b2a6ae55315f" (UID: "bab95882-60a4-4e02-b3b7-b2a6ae55315f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.923602 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-config-data" (OuterVolumeSpecName: "config-data") pod "bab95882-60a4-4e02-b3b7-b2a6ae55315f" (UID: "bab95882-60a4-4e02-b3b7-b2a6ae55315f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.979634 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:55 crc kubenswrapper[4894]: I1209 15:49:55.979662 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bab95882-60a4-4e02-b3b7-b2a6ae55315f-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.044700 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.046454 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5776b569cc-b4c82" event={"ID":"7a921fbc-ca69-439e-b357-eb15cc4de86e","Type":"ContainerStarted","Data":"13e56d3e90de2b354fd90a37be349280f5ce31ff63e1c333bee4d3c554241710"} Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.053330 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-569f9c8ccb-pkfj6" event={"ID":"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992","Type":"ContainerStarted","Data":"11ee86bb97843f66a3b73c31e5f991932bfb9b69347db6f215e2c4493afc94fe"} Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.053382 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-569f9c8ccb-pkfj6" event={"ID":"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992","Type":"ContainerStarted","Data":"4d4aed6ba4c61e4e05b19fe908160be8df98cfde903dbb0f43021286e99abe42"} Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.055151 4894 generic.go:334] "Generic (PLEG): container finished" podID="fee9920a-8d2a-4bf3-88dc-5d745c0dee82" containerID="022b1ebe3b3629d7205565091075b56883dc4dad0363843d288d355963424db4" exitCode=0 Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.055210 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75479fdbd9-z4s7v" event={"ID":"fee9920a-8d2a-4bf3-88dc-5d745c0dee82","Type":"ContainerDied","Data":"022b1ebe3b3629d7205565091075b56883dc4dad0363843d288d355963424db4"} Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.055231 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75479fdbd9-z4s7v" event={"ID":"fee9920a-8d2a-4bf3-88dc-5d745c0dee82","Type":"ContainerStarted","Data":"ae4d186a57c63f678215acc088926fa2856567f4b8f325c85444e18740e96676"} Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.062903 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.062999 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"bab95882-60a4-4e02-b3b7-b2a6ae55315f","Type":"ContainerDied","Data":"7ff237304503b28262d2010661c11d69f410f1f9f57bdf215964ba305c37ef41"} Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.063053 4894 scope.go:117] "RemoveContainer" containerID="2a84f533d78019c9f713e6d77e8cd58d1ca1eee10176c65b3466d16f959bb39e" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.064425 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-67f4558b8d-v6cpx" event={"ID":"b470d89f-2e08-4222-adc9-5dab929008ea","Type":"ContainerStarted","Data":"fd3b7c6691ce6606c3a0a449d9ac31cbc3d861c09a191c99074cd4d8e0296a14"} Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.094414 4894 scope.go:117] "RemoveContainer" containerID="36c3822c1e6a90d261e05cba628235eab698bf1a47839de9a35bce0fe9dc2b06" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.172978 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.204924 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.220129 4894 scope.go:117] "RemoveContainer" containerID="ad993b83ca0bae3b9d2a6aa998c2747bd8ca7896fb6279c3e80ae74e1c99bd1f" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.252845 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:49:56 crc kubenswrapper[4894]: E1209 15:49:56.253560 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bab95882-60a4-4e02-b3b7-b2a6ae55315f" containerName="ceilometer-notification-agent" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.253679 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="bab95882-60a4-4e02-b3b7-b2a6ae55315f" containerName="ceilometer-notification-agent" Dec 09 15:49:56 crc kubenswrapper[4894]: E1209 15:49:56.253742 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bab95882-60a4-4e02-b3b7-b2a6ae55315f" containerName="ceilometer-central-agent" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.253753 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="bab95882-60a4-4e02-b3b7-b2a6ae55315f" containerName="ceilometer-central-agent" Dec 09 15:49:56 crc kubenswrapper[4894]: E1209 15:49:56.253777 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bab95882-60a4-4e02-b3b7-b2a6ae55315f" containerName="proxy-httpd" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.253786 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="bab95882-60a4-4e02-b3b7-b2a6ae55315f" containerName="proxy-httpd" Dec 09 15:49:56 crc kubenswrapper[4894]: E1209 15:49:56.253854 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bab95882-60a4-4e02-b3b7-b2a6ae55315f" containerName="sg-core" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.253865 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="bab95882-60a4-4e02-b3b7-b2a6ae55315f" containerName="sg-core" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.254105 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="bab95882-60a4-4e02-b3b7-b2a6ae55315f" containerName="proxy-httpd" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.254163 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="bab95882-60a4-4e02-b3b7-b2a6ae55315f" containerName="ceilometer-central-agent" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.254177 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="bab95882-60a4-4e02-b3b7-b2a6ae55315f" containerName="sg-core" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.254200 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="bab95882-60a4-4e02-b3b7-b2a6ae55315f" containerName="ceilometer-notification-agent" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.256416 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.259901 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.261072 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.265131 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.281855 4894 scope.go:117] "RemoveContainer" containerID="42832b08996b48d3dee186127faf58ac23a8ebe1bcccbb9bc653bbc535ff663c" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.289237 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.289267 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-scripts\") pod \"ceilometer-0\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.289693 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ts7pt\" (UniqueName: \"kubernetes.io/projected/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-kube-api-access-ts7pt\") pod \"ceilometer-0\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.289760 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-config-data\") pod \"ceilometer-0\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.289799 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.289829 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-log-httpd\") pod \"ceilometer-0\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.289893 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-run-httpd\") pod \"ceilometer-0\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.307073 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.356121 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-68c677b759-g4lpm"] Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.400744 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ts7pt\" (UniqueName: \"kubernetes.io/projected/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-kube-api-access-ts7pt\") pod \"ceilometer-0\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.400798 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-config-data\") pod \"ceilometer-0\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.400819 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.400840 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-log-httpd\") pod \"ceilometer-0\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.400871 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-run-httpd\") pod \"ceilometer-0\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.400935 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.400956 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-scripts\") pod \"ceilometer-0\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.407931 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-run-httpd\") pod \"ceilometer-0\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.408201 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-log-httpd\") pod \"ceilometer-0\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.420384 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-config-data\") pod \"ceilometer-0\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.446799 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.477997 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-scripts\") pod \"ceilometer-0\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.486504 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.488320 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ts7pt\" (UniqueName: \"kubernetes.io/projected/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-kube-api-access-ts7pt\") pod \"ceilometer-0\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.598331 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.666510 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75479fdbd9-z4s7v" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.709745 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-ovsdbserver-nb\") pod \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\" (UID: \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\") " Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.710068 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mbm4g\" (UniqueName: \"kubernetes.io/projected/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-kube-api-access-mbm4g\") pod \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\" (UID: \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\") " Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.710098 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-ovsdbserver-sb\") pod \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\" (UID: \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\") " Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.710127 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-config\") pod \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\" (UID: \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\") " Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.710175 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-dns-svc\") pod \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\" (UID: \"fee9920a-8d2a-4bf3-88dc-5d745c0dee82\") " Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.722799 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-kube-api-access-mbm4g" (OuterVolumeSpecName: "kube-api-access-mbm4g") pod "fee9920a-8d2a-4bf3-88dc-5d745c0dee82" (UID: "fee9920a-8d2a-4bf3-88dc-5d745c0dee82"). InnerVolumeSpecName "kube-api-access-mbm4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.772965 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.812193 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mbm4g\" (UniqueName: \"kubernetes.io/projected/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-kube-api-access-mbm4g\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.831520 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "fee9920a-8d2a-4bf3-88dc-5d745c0dee82" (UID: "fee9920a-8d2a-4bf3-88dc-5d745c0dee82"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.832801 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-config" (OuterVolumeSpecName: "config") pod "fee9920a-8d2a-4bf3-88dc-5d745c0dee82" (UID: "fee9920a-8d2a-4bf3-88dc-5d745c0dee82"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.844369 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "fee9920a-8d2a-4bf3-88dc-5d745c0dee82" (UID: "fee9920a-8d2a-4bf3-88dc-5d745c0dee82"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.845152 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "fee9920a-8d2a-4bf3-88dc-5d745c0dee82" (UID: "fee9920a-8d2a-4bf3-88dc-5d745c0dee82"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.913559 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.913593 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.913604 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:56 crc kubenswrapper[4894]: I1209 15:49:56.913614 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/fee9920a-8d2a-4bf3-88dc-5d745c0dee82-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 15:49:57 crc kubenswrapper[4894]: I1209 15:49:57.073769 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-75479fdbd9-z4s7v" event={"ID":"fee9920a-8d2a-4bf3-88dc-5d745c0dee82","Type":"ContainerDied","Data":"ae4d186a57c63f678215acc088926fa2856567f4b8f325c85444e18740e96676"} Dec 09 15:49:57 crc kubenswrapper[4894]: I1209 15:49:57.073841 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-75479fdbd9-z4s7v" Dec 09 15:49:57 crc kubenswrapper[4894]: I1209 15:49:57.074129 4894 scope.go:117] "RemoveContainer" containerID="022b1ebe3b3629d7205565091075b56883dc4dad0363843d288d355963424db4" Dec 09 15:49:57 crc kubenswrapper[4894]: I1209 15:49:57.075609 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f8c380b2-49de-4630-80e9-15c73cd99045","Type":"ContainerStarted","Data":"f8aeec9de6975d944e48cedf195b0b7cb434ee40bcda5accecffffb3d8f4bd59"} Dec 09 15:49:57 crc kubenswrapper[4894]: I1209 15:49:57.079489 4894 generic.go:334] "Generic (PLEG): container finished" podID="d3f62001-a311-45d1-ab65-15cd3de84d4e" containerID="877dd577395d8d411177a9ca04183dcb4a4add06bd2ce7a00dd54e8902046404" exitCode=0 Dec 09 15:49:57 crc kubenswrapper[4894]: I1209 15:49:57.079563 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68c677b759-g4lpm" event={"ID":"d3f62001-a311-45d1-ab65-15cd3de84d4e","Type":"ContainerDied","Data":"877dd577395d8d411177a9ca04183dcb4a4add06bd2ce7a00dd54e8902046404"} Dec 09 15:49:57 crc kubenswrapper[4894]: I1209 15:49:57.079591 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68c677b759-g4lpm" event={"ID":"d3f62001-a311-45d1-ab65-15cd3de84d4e","Type":"ContainerStarted","Data":"0f422df104602f440c514ebe0b707da27dafca695e003ca2d519c8d6bbb7e39f"} Dec 09 15:49:57 crc kubenswrapper[4894]: I1209 15:49:57.084060 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-569f9c8ccb-pkfj6" event={"ID":"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992","Type":"ContainerStarted","Data":"ed959f038ef472ffa42f39eb22c506846d5127725413646ef9b8a0ecbf5c25be"} Dec 09 15:49:57 crc kubenswrapper[4894]: I1209 15:49:57.084908 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:49:57 crc kubenswrapper[4894]: I1209 15:49:57.084938 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:49:57 crc kubenswrapper[4894]: I1209 15:49:57.162306 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-569f9c8ccb-pkfj6" podStartSLOduration=3.16228686 podStartE2EDuration="3.16228686s" podCreationTimestamp="2025-12-09 15:49:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:49:57.125761025 +0000 UTC m=+1091.444971694" watchObservedRunningTime="2025-12-09 15:49:57.16228686 +0000 UTC m=+1091.481497529" Dec 09 15:49:57 crc kubenswrapper[4894]: I1209 15:49:57.181350 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-75479fdbd9-z4s7v"] Dec 09 15:49:57 crc kubenswrapper[4894]: I1209 15:49:57.181411 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-75479fdbd9-z4s7v"] Dec 09 15:49:57 crc kubenswrapper[4894]: I1209 15:49:57.885033 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 09 15:49:58 crc kubenswrapper[4894]: I1209 15:49:58.073722 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:49:58 crc kubenswrapper[4894]: W1209 15:49:58.097345 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2e1922e1_492f_4bee_8b4c_eef4f06eeae2.slice/crio-f1ff300d87ab8014e68a5216b78a266a3cacab1ea59b1417d7209c1731767b41 WatchSource:0}: Error finding container f1ff300d87ab8014e68a5216b78a266a3cacab1ea59b1417d7209c1731767b41: Status 404 returned error can't find the container with id f1ff300d87ab8014e68a5216b78a266a3cacab1ea59b1417d7209c1731767b41 Dec 09 15:49:58 crc kubenswrapper[4894]: I1209 15:49:58.101844 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e04dd18f-5e7f-464b-9f65-d4233a0753a6","Type":"ContainerStarted","Data":"b9165e1c3bb9bf6c3e66ea837c1fd3779dec0716ef9889854264ba19259d88f5"} Dec 09 15:49:58 crc kubenswrapper[4894]: I1209 15:49:58.133136 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bab95882-60a4-4e02-b3b7-b2a6ae55315f" path="/var/lib/kubelet/pods/bab95882-60a4-4e02-b3b7-b2a6ae55315f/volumes" Dec 09 15:49:58 crc kubenswrapper[4894]: I1209 15:49:58.134558 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fee9920a-8d2a-4bf3-88dc-5d745c0dee82" path="/var/lib/kubelet/pods/fee9920a-8d2a-4bf3-88dc-5d745c0dee82/volumes" Dec 09 15:49:58 crc kubenswrapper[4894]: I1209 15:49:58.135193 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-68c677b759-g4lpm" Dec 09 15:49:58 crc kubenswrapper[4894]: I1209 15:49:58.135958 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68c677b759-g4lpm" event={"ID":"d3f62001-a311-45d1-ab65-15cd3de84d4e","Type":"ContainerStarted","Data":"038221b6bd6c60ebdac98f041b56b3384ae28d03a6a88aaee2fad23b522e57b7"} Dec 09 15:49:58 crc kubenswrapper[4894]: I1209 15:49:58.135983 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-67f4558b8d-v6cpx" event={"ID":"b470d89f-2e08-4222-adc9-5dab929008ea","Type":"ContainerStarted","Data":"81036da6f7f0028990707f82180b6ee01dba28bfcca6cb8edfa6ae40fe44ab96"} Dec 09 15:49:58 crc kubenswrapper[4894]: I1209 15:49:58.136055 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5776b569cc-b4c82" event={"ID":"7a921fbc-ca69-439e-b357-eb15cc4de86e","Type":"ContainerStarted","Data":"2fa4060a059a2b0420fdaf0415adfff6c40be08deed07083b963730db5c19970"} Dec 09 15:49:58 crc kubenswrapper[4894]: I1209 15:49:58.147046 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-68c677b759-g4lpm" podStartSLOduration=3.147027687 podStartE2EDuration="3.147027687s" podCreationTimestamp="2025-12-09 15:49:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:49:58.140779963 +0000 UTC m=+1092.459990642" watchObservedRunningTime="2025-12-09 15:49:58.147027687 +0000 UTC m=+1092.466238356" Dec 09 15:49:59 crc kubenswrapper[4894]: I1209 15:49:59.150016 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e04dd18f-5e7f-464b-9f65-d4233a0753a6","Type":"ContainerStarted","Data":"bc026d7256867bfaa5ff971d4905f1cb9f84d1c85773c30357b866f8c020b1a6"} Dec 09 15:49:59 crc kubenswrapper[4894]: I1209 15:49:59.154049 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-67f4558b8d-v6cpx" event={"ID":"b470d89f-2e08-4222-adc9-5dab929008ea","Type":"ContainerStarted","Data":"52523e2f9d2cf45fd1d2545324b222b596c39e02dfdc039d69ccf59675d3e8c3"} Dec 09 15:49:59 crc kubenswrapper[4894]: I1209 15:49:59.156044 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-5776b569cc-b4c82" event={"ID":"7a921fbc-ca69-439e-b357-eb15cc4de86e","Type":"ContainerStarted","Data":"2ba5db439caba0abdfda6195e8bddee48793cb90531cd6c77f94451e26212f30"} Dec 09 15:49:59 crc kubenswrapper[4894]: I1209 15:49:59.158816 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e1922e1-492f-4bee-8b4c-eef4f06eeae2","Type":"ContainerStarted","Data":"9f5054fa5f3cb566f03ee591a1ff9f0d5101dc1d2a72a750987d27b7c844b785"} Dec 09 15:49:59 crc kubenswrapper[4894]: I1209 15:49:59.158879 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e1922e1-492f-4bee-8b4c-eef4f06eeae2","Type":"ContainerStarted","Data":"f1ff300d87ab8014e68a5216b78a266a3cacab1ea59b1417d7209c1731767b41"} Dec 09 15:49:59 crc kubenswrapper[4894]: I1209 15:49:59.165002 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f8c380b2-49de-4630-80e9-15c73cd99045","Type":"ContainerStarted","Data":"cd3e93d7b9ee1327ddfcd05f826c2a2212772a728777956cc2706bfa377b377d"} Dec 09 15:49:59 crc kubenswrapper[4894]: I1209 15:49:59.196819 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-67f4558b8d-v6cpx" podStartSLOduration=3.016773217 podStartE2EDuration="5.196801574s" podCreationTimestamp="2025-12-09 15:49:54 +0000 UTC" firstStartedPulling="2025-12-09 15:49:55.549576921 +0000 UTC m=+1089.868787590" lastFinishedPulling="2025-12-09 15:49:57.729605288 +0000 UTC m=+1092.048815947" observedRunningTime="2025-12-09 15:49:59.173205377 +0000 UTC m=+1093.492416046" watchObservedRunningTime="2025-12-09 15:49:59.196801574 +0000 UTC m=+1093.516012243" Dec 09 15:49:59 crc kubenswrapper[4894]: I1209 15:49:59.199617 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-5776b569cc-b4c82" podStartSLOduration=2.602526291 podStartE2EDuration="5.199607378s" podCreationTimestamp="2025-12-09 15:49:54 +0000 UTC" firstStartedPulling="2025-12-09 15:49:55.056838323 +0000 UTC m=+1089.376048992" lastFinishedPulling="2025-12-09 15:49:57.65391941 +0000 UTC m=+1091.973130079" observedRunningTime="2025-12-09 15:49:59.196181787 +0000 UTC m=+1093.515392456" watchObservedRunningTime="2025-12-09 15:49:59.199607378 +0000 UTC m=+1093.518818037" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.175552 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f8c380b2-49de-4630-80e9-15c73cd99045","Type":"ContainerStarted","Data":"b3e7d4a21768e3a06683a4ec83b56a18c03e4d8da51242eb72a21e484736b54d"} Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.178042 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e1922e1-492f-4bee-8b4c-eef4f06eeae2","Type":"ContainerStarted","Data":"1b6b37c96e9bec2f6472cee4d4e32416786dd67b4f3dd297845e0fc86137be2f"} Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.180712 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e04dd18f-5e7f-464b-9f65-d4233a0753a6" containerName="cinder-api-log" containerID="cri-o://bc026d7256867bfaa5ff971d4905f1cb9f84d1c85773c30357b866f8c020b1a6" gracePeriod=30 Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.181309 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e04dd18f-5e7f-464b-9f65-d4233a0753a6","Type":"ContainerStarted","Data":"d40ea04a7f20b471d1866ee30e226d525549f3c69801618a9f0f0923e28b2a40"} Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.181573 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.182244 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="e04dd18f-5e7f-464b-9f65-d4233a0753a6" containerName="cinder-api" containerID="cri-o://d40ea04a7f20b471d1866ee30e226d525549f3c69801618a9f0f0923e28b2a40" gracePeriod=30 Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.213956 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.766267062 podStartE2EDuration="5.213929847s" podCreationTimestamp="2025-12-09 15:49:55 +0000 UTC" firstStartedPulling="2025-12-09 15:49:56.288857723 +0000 UTC m=+1090.608068392" lastFinishedPulling="2025-12-09 15:49:57.736520508 +0000 UTC m=+1092.055731177" observedRunningTime="2025-12-09 15:50:00.198228467 +0000 UTC m=+1094.517439146" watchObservedRunningTime="2025-12-09 15:50:00.213929847 +0000 UTC m=+1094.533140556" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.241809 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=5.241784435 podStartE2EDuration="5.241784435s" podCreationTimestamp="2025-12-09 15:49:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:50:00.220735665 +0000 UTC m=+1094.539946334" watchObservedRunningTime="2025-12-09 15:50:00.241784435 +0000 UTC m=+1094.560995124" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.458599 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.645441 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.845427 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-5df7cc9c4d-kfjtd"] Dec 09 15:50:00 crc kubenswrapper[4894]: E1209 15:50:00.845783 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fee9920a-8d2a-4bf3-88dc-5d745c0dee82" containerName="init" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.845799 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fee9920a-8d2a-4bf3-88dc-5d745c0dee82" containerName="init" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.845993 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fee9920a-8d2a-4bf3-88dc-5d745c0dee82" containerName="init" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.846888 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.850522 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.850614 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.865488 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5df7cc9c4d-kfjtd"] Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.883513 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcc16386-61b0-4e99-a220-70f90f549e90-config-data-custom\") pod \"barbican-api-5df7cc9c4d-kfjtd\" (UID: \"fcc16386-61b0-4e99-a220-70f90f549e90\") " pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.883560 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fcc16386-61b0-4e99-a220-70f90f549e90-internal-tls-certs\") pod \"barbican-api-5df7cc9c4d-kfjtd\" (UID: \"fcc16386-61b0-4e99-a220-70f90f549e90\") " pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.883592 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f6zmz\" (UniqueName: \"kubernetes.io/projected/fcc16386-61b0-4e99-a220-70f90f549e90-kube-api-access-f6zmz\") pod \"barbican-api-5df7cc9c4d-kfjtd\" (UID: \"fcc16386-61b0-4e99-a220-70f90f549e90\") " pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.883683 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fcc16386-61b0-4e99-a220-70f90f549e90-public-tls-certs\") pod \"barbican-api-5df7cc9c4d-kfjtd\" (UID: \"fcc16386-61b0-4e99-a220-70f90f549e90\") " pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.883758 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcc16386-61b0-4e99-a220-70f90f549e90-logs\") pod \"barbican-api-5df7cc9c4d-kfjtd\" (UID: \"fcc16386-61b0-4e99-a220-70f90f549e90\") " pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.883775 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcc16386-61b0-4e99-a220-70f90f549e90-config-data\") pod \"barbican-api-5df7cc9c4d-kfjtd\" (UID: \"fcc16386-61b0-4e99-a220-70f90f549e90\") " pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.883789 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcc16386-61b0-4e99-a220-70f90f549e90-combined-ca-bundle\") pod \"barbican-api-5df7cc9c4d-kfjtd\" (UID: \"fcc16386-61b0-4e99-a220-70f90f549e90\") " pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.985511 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcc16386-61b0-4e99-a220-70f90f549e90-logs\") pod \"barbican-api-5df7cc9c4d-kfjtd\" (UID: \"fcc16386-61b0-4e99-a220-70f90f549e90\") " pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.985555 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcc16386-61b0-4e99-a220-70f90f549e90-config-data\") pod \"barbican-api-5df7cc9c4d-kfjtd\" (UID: \"fcc16386-61b0-4e99-a220-70f90f549e90\") " pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.985577 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcc16386-61b0-4e99-a220-70f90f549e90-combined-ca-bundle\") pod \"barbican-api-5df7cc9c4d-kfjtd\" (UID: \"fcc16386-61b0-4e99-a220-70f90f549e90\") " pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.985643 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcc16386-61b0-4e99-a220-70f90f549e90-config-data-custom\") pod \"barbican-api-5df7cc9c4d-kfjtd\" (UID: \"fcc16386-61b0-4e99-a220-70f90f549e90\") " pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.985674 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fcc16386-61b0-4e99-a220-70f90f549e90-internal-tls-certs\") pod \"barbican-api-5df7cc9c4d-kfjtd\" (UID: \"fcc16386-61b0-4e99-a220-70f90f549e90\") " pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.985697 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f6zmz\" (UniqueName: \"kubernetes.io/projected/fcc16386-61b0-4e99-a220-70f90f549e90-kube-api-access-f6zmz\") pod \"barbican-api-5df7cc9c4d-kfjtd\" (UID: \"fcc16386-61b0-4e99-a220-70f90f549e90\") " pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.985742 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fcc16386-61b0-4e99-a220-70f90f549e90-public-tls-certs\") pod \"barbican-api-5df7cc9c4d-kfjtd\" (UID: \"fcc16386-61b0-4e99-a220-70f90f549e90\") " pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.986685 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fcc16386-61b0-4e99-a220-70f90f549e90-logs\") pod \"barbican-api-5df7cc9c4d-kfjtd\" (UID: \"fcc16386-61b0-4e99-a220-70f90f549e90\") " pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.992740 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fcc16386-61b0-4e99-a220-70f90f549e90-config-data\") pod \"barbican-api-5df7cc9c4d-kfjtd\" (UID: \"fcc16386-61b0-4e99-a220-70f90f549e90\") " pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.993206 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fcc16386-61b0-4e99-a220-70f90f549e90-config-data-custom\") pod \"barbican-api-5df7cc9c4d-kfjtd\" (UID: \"fcc16386-61b0-4e99-a220-70f90f549e90\") " pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.993767 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fcc16386-61b0-4e99-a220-70f90f549e90-combined-ca-bundle\") pod \"barbican-api-5df7cc9c4d-kfjtd\" (UID: \"fcc16386-61b0-4e99-a220-70f90f549e90\") " pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:00 crc kubenswrapper[4894]: I1209 15:50:00.994474 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fcc16386-61b0-4e99-a220-70f90f549e90-internal-tls-certs\") pod \"barbican-api-5df7cc9c4d-kfjtd\" (UID: \"fcc16386-61b0-4e99-a220-70f90f549e90\") " pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:01 crc kubenswrapper[4894]: I1209 15:50:01.001344 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fcc16386-61b0-4e99-a220-70f90f549e90-public-tls-certs\") pod \"barbican-api-5df7cc9c4d-kfjtd\" (UID: \"fcc16386-61b0-4e99-a220-70f90f549e90\") " pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:01 crc kubenswrapper[4894]: I1209 15:50:01.004104 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:50:01 crc kubenswrapper[4894]: I1209 15:50:01.009615 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f6zmz\" (UniqueName: \"kubernetes.io/projected/fcc16386-61b0-4e99-a220-70f90f549e90-kube-api-access-f6zmz\") pod \"barbican-api-5df7cc9c4d-kfjtd\" (UID: \"fcc16386-61b0-4e99-a220-70f90f549e90\") " pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:01 crc kubenswrapper[4894]: I1209 15:50:01.190876 4894 generic.go:334] "Generic (PLEG): container finished" podID="e04dd18f-5e7f-464b-9f65-d4233a0753a6" containerID="bc026d7256867bfaa5ff971d4905f1cb9f84d1c85773c30357b866f8c020b1a6" exitCode=143 Dec 09 15:50:01 crc kubenswrapper[4894]: I1209 15:50:01.190962 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e04dd18f-5e7f-464b-9f65-d4233a0753a6","Type":"ContainerDied","Data":"bc026d7256867bfaa5ff971d4905f1cb9f84d1c85773c30357b866f8c020b1a6"} Dec 09 15:50:01 crc kubenswrapper[4894]: I1209 15:50:01.210967 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:01 crc kubenswrapper[4894]: I1209 15:50:01.911135 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-5df7cc9c4d-kfjtd"] Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.202864 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5df7cc9c4d-kfjtd" event={"ID":"fcc16386-61b0-4e99-a220-70f90f549e90","Type":"ContainerStarted","Data":"5231761dc5c99b3f9a9cc3947d9e16f82f92c11983c8c6910060ffc7a373681e"} Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.218815 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e1922e1-492f-4bee-8b4c-eef4f06eeae2","Type":"ContainerStarted","Data":"4e84eada7e930987ecb2f68c26a9194295aa68b506792299af08d2ba022cd9c2"} Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.221772 4894 generic.go:334] "Generic (PLEG): container finished" podID="e04dd18f-5e7f-464b-9f65-d4233a0753a6" containerID="d40ea04a7f20b471d1866ee30e226d525549f3c69801618a9f0f0923e28b2a40" exitCode=0 Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.223694 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e04dd18f-5e7f-464b-9f65-d4233a0753a6","Type":"ContainerDied","Data":"d40ea04a7f20b471d1866ee30e226d525549f3c69801618a9f0f0923e28b2a40"} Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.336414 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.662565 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.825815 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e04dd18f-5e7f-464b-9f65-d4233a0753a6-logs\") pod \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.825894 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-config-data-custom\") pod \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.825914 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6n4vb\" (UniqueName: \"kubernetes.io/projected/e04dd18f-5e7f-464b-9f65-d4233a0753a6-kube-api-access-6n4vb\") pod \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.825951 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-config-data\") pod \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.825986 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-combined-ca-bundle\") pod \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.826050 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-scripts\") pod \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.826113 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e04dd18f-5e7f-464b-9f65-d4233a0753a6-etc-machine-id\") pod \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\" (UID: \"e04dd18f-5e7f-464b-9f65-d4233a0753a6\") " Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.826492 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e04dd18f-5e7f-464b-9f65-d4233a0753a6-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "e04dd18f-5e7f-464b-9f65-d4233a0753a6" (UID: "e04dd18f-5e7f-464b-9f65-d4233a0753a6"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.827571 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e04dd18f-5e7f-464b-9f65-d4233a0753a6-logs" (OuterVolumeSpecName: "logs") pod "e04dd18f-5e7f-464b-9f65-d4233a0753a6" (UID: "e04dd18f-5e7f-464b-9f65-d4233a0753a6"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.831687 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e04dd18f-5e7f-464b-9f65-d4233a0753a6-kube-api-access-6n4vb" (OuterVolumeSpecName: "kube-api-access-6n4vb") pod "e04dd18f-5e7f-464b-9f65-d4233a0753a6" (UID: "e04dd18f-5e7f-464b-9f65-d4233a0753a6"). InnerVolumeSpecName "kube-api-access-6n4vb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.831974 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "e04dd18f-5e7f-464b-9f65-d4233a0753a6" (UID: "e04dd18f-5e7f-464b-9f65-d4233a0753a6"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.832514 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-scripts" (OuterVolumeSpecName: "scripts") pod "e04dd18f-5e7f-464b-9f65-d4233a0753a6" (UID: "e04dd18f-5e7f-464b-9f65-d4233a0753a6"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.854142 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e04dd18f-5e7f-464b-9f65-d4233a0753a6" (UID: "e04dd18f-5e7f-464b-9f65-d4233a0753a6"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.878158 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-config-data" (OuterVolumeSpecName: "config-data") pod "e04dd18f-5e7f-464b-9f65-d4233a0753a6" (UID: "e04dd18f-5e7f-464b-9f65-d4233a0753a6"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.927823 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e04dd18f-5e7f-464b-9f65-d4233a0753a6-logs\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.928014 4894 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.928071 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6n4vb\" (UniqueName: \"kubernetes.io/projected/e04dd18f-5e7f-464b-9f65-d4233a0753a6-kube-api-access-6n4vb\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.928127 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.929511 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.929550 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e04dd18f-5e7f-464b-9f65-d4233a0753a6-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.929560 4894 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/e04dd18f-5e7f-464b-9f65-d4233a0753a6-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:02 crc kubenswrapper[4894]: I1209 15:50:02.962986 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/horizon-6b9bcbff6b-l6n7z" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.029357 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-56f65bfbb-tv5ld"] Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.232510 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"e04dd18f-5e7f-464b-9f65-d4233a0753a6","Type":"ContainerDied","Data":"b9165e1c3bb9bf6c3e66ea837c1fd3779dec0716ef9889854264ba19259d88f5"} Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.232559 4894 scope.go:117] "RemoveContainer" containerID="d40ea04a7f20b471d1866ee30e226d525549f3c69801618a9f0f0923e28b2a40" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.232694 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.244066 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-56f65bfbb-tv5ld" podUID="0e41ac7e-15c8-4a9e-ae96-2e68b055ea16" containerName="horizon-log" containerID="cri-o://b2f96b6acf368b01f8d4a3be9093d23033af84a3075e278756e462ebdc719a8b" gracePeriod=30 Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.244095 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5df7cc9c4d-kfjtd" event={"ID":"fcc16386-61b0-4e99-a220-70f90f549e90","Type":"ContainerStarted","Data":"94e1c39fbe26f7727f80e59591d79c6ad5906a862e03f09474b93096a6f3b514"} Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.244138 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-5df7cc9c4d-kfjtd" event={"ID":"fcc16386-61b0-4e99-a220-70f90f549e90","Type":"ContainerStarted","Data":"9257b1fc4d076e1d63301476524d1f25fcdff16ed46c068267f25a212809682b"} Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.244285 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.244347 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/horizon-56f65bfbb-tv5ld" podUID="0e41ac7e-15c8-4a9e-ae96-2e68b055ea16" containerName="horizon" containerID="cri-o://1b3e955a174036764325c570292b5dde1384ef1df60f2868c564915e82f08177" gracePeriod=30 Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.244983 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.294318 4894 scope.go:117] "RemoveContainer" containerID="bc026d7256867bfaa5ff971d4905f1cb9f84d1c85773c30357b866f8c020b1a6" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.306590 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-5df7cc9c4d-kfjtd" podStartSLOduration=3.306570116 podStartE2EDuration="3.306570116s" podCreationTimestamp="2025-12-09 15:50:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:50:03.283107533 +0000 UTC m=+1097.602318212" watchObservedRunningTime="2025-12-09 15:50:03.306570116 +0000 UTC m=+1097.625780785" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.314946 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.344366 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.359823 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Dec 09 15:50:03 crc kubenswrapper[4894]: E1209 15:50:03.360174 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e04dd18f-5e7f-464b-9f65-d4233a0753a6" containerName="cinder-api-log" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.360190 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e04dd18f-5e7f-464b-9f65-d4233a0753a6" containerName="cinder-api-log" Dec 09 15:50:03 crc kubenswrapper[4894]: E1209 15:50:03.360211 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e04dd18f-5e7f-464b-9f65-d4233a0753a6" containerName="cinder-api" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.360219 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e04dd18f-5e7f-464b-9f65-d4233a0753a6" containerName="cinder-api" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.360385 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e04dd18f-5e7f-464b-9f65-d4233a0753a6" containerName="cinder-api" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.360400 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e04dd18f-5e7f-464b-9f65-d4233a0753a6" containerName="cinder-api-log" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.368396 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.372177 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.372358 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.372465 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.382753 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.443545 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/58e3314f-76b0-4e83-93ce-39c342c44acb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.443599 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58e3314f-76b0-4e83-93ce-39c342c44acb-logs\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.443660 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58e3314f-76b0-4e83-93ce-39c342c44acb-config-data-custom\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.443687 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58e3314f-76b0-4e83-93ce-39c342c44acb-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.443751 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/58e3314f-76b0-4e83-93ce-39c342c44acb-public-tls-certs\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.443767 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58e3314f-76b0-4e83-93ce-39c342c44acb-scripts\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.443792 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g72gt\" (UniqueName: \"kubernetes.io/projected/58e3314f-76b0-4e83-93ce-39c342c44acb-kube-api-access-g72gt\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.443850 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58e3314f-76b0-4e83-93ce-39c342c44acb-config-data\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.443878 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e3314f-76b0-4e83-93ce-39c342c44acb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.544390 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58e3314f-76b0-4e83-93ce-39c342c44acb-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.544460 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/58e3314f-76b0-4e83-93ce-39c342c44acb-public-tls-certs\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.544482 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58e3314f-76b0-4e83-93ce-39c342c44acb-scripts\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.544504 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g72gt\" (UniqueName: \"kubernetes.io/projected/58e3314f-76b0-4e83-93ce-39c342c44acb-kube-api-access-g72gt\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.544557 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58e3314f-76b0-4e83-93ce-39c342c44acb-config-data\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.544624 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e3314f-76b0-4e83-93ce-39c342c44acb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.544682 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/58e3314f-76b0-4e83-93ce-39c342c44acb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.544712 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58e3314f-76b0-4e83-93ce-39c342c44acb-logs\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.545271 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/58e3314f-76b0-4e83-93ce-39c342c44acb-logs\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.545364 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58e3314f-76b0-4e83-93ce-39c342c44acb-config-data-custom\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.546197 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/58e3314f-76b0-4e83-93ce-39c342c44acb-etc-machine-id\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.549394 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/58e3314f-76b0-4e83-93ce-39c342c44acb-scripts\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.549885 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/58e3314f-76b0-4e83-93ce-39c342c44acb-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.550113 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/58e3314f-76b0-4e83-93ce-39c342c44acb-public-tls-certs\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.550970 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/58e3314f-76b0-4e83-93ce-39c342c44acb-config-data\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.551171 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/58e3314f-76b0-4e83-93ce-39c342c44acb-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.551680 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/58e3314f-76b0-4e83-93ce-39c342c44acb-config-data-custom\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.575060 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g72gt\" (UniqueName: \"kubernetes.io/projected/58e3314f-76b0-4e83-93ce-39c342c44acb-kube-api-access-g72gt\") pod \"cinder-api-0\" (UID: \"58e3314f-76b0-4e83-93ce-39c342c44acb\") " pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.698017 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.938839 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-56fb46ccd4-rwfjk" Dec 09 15:50:03 crc kubenswrapper[4894]: I1209 15:50:03.956869 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-56fb46ccd4-rwfjk" Dec 09 15:50:04 crc kubenswrapper[4894]: I1209 15:50:04.122980 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e04dd18f-5e7f-464b-9f65-d4233a0753a6" path="/var/lib/kubelet/pods/e04dd18f-5e7f-464b-9f65-d4233a0753a6/volumes" Dec 09 15:50:04 crc kubenswrapper[4894]: W1209 15:50:04.155798 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod58e3314f_76b0_4e83_93ce_39c342c44acb.slice/crio-a09c9f0d4e9e48cfcfd838f433739e49b3b268824669261a85a0ab52ec99c60d WatchSource:0}: Error finding container a09c9f0d4e9e48cfcfd838f433739e49b3b268824669261a85a0ab52ec99c60d: Status 404 returned error can't find the container with id a09c9f0d4e9e48cfcfd838f433739e49b3b268824669261a85a0ab52ec99c60d Dec 09 15:50:04 crc kubenswrapper[4894]: I1209 15:50:04.159723 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Dec 09 15:50:04 crc kubenswrapper[4894]: I1209 15:50:04.270793 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e1922e1-492f-4bee-8b4c-eef4f06eeae2","Type":"ContainerStarted","Data":"971e652bbce0431e8968ac6720e5444c04db44ea13f7178b12fa57b109401870"} Dec 09 15:50:04 crc kubenswrapper[4894]: I1209 15:50:04.271149 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 15:50:04 crc kubenswrapper[4894]: I1209 15:50:04.276240 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"58e3314f-76b0-4e83-93ce-39c342c44acb","Type":"ContainerStarted","Data":"a09c9f0d4e9e48cfcfd838f433739e49b3b268824669261a85a0ab52ec99c60d"} Dec 09 15:50:04 crc kubenswrapper[4894]: I1209 15:50:04.297116 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.456591788 podStartE2EDuration="8.297095964s" podCreationTimestamp="2025-12-09 15:49:56 +0000 UTC" firstStartedPulling="2025-12-09 15:49:58.106002055 +0000 UTC m=+1092.425212714" lastFinishedPulling="2025-12-09 15:50:03.946506221 +0000 UTC m=+1098.265716890" observedRunningTime="2025-12-09 15:50:04.293562532 +0000 UTC m=+1098.612773201" watchObservedRunningTime="2025-12-09 15:50:04.297095964 +0000 UTC m=+1098.616306633" Dec 09 15:50:05 crc kubenswrapper[4894]: I1209 15:50:05.287503 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"58e3314f-76b0-4e83-93ce-39c342c44acb","Type":"ContainerStarted","Data":"6f263d7914e24d17b55893e485acba54d823cceb4b5f20ef9506fe01bfc68e05"} Dec 09 15:50:05 crc kubenswrapper[4894]: I1209 15:50:05.683445 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-68c677b759-g4lpm" Dec 09 15:50:05 crc kubenswrapper[4894]: I1209 15:50:05.740017 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f69459849-bjgnn"] Dec 09 15:50:05 crc kubenswrapper[4894]: I1209 15:50:05.740332 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7f69459849-bjgnn" podUID="7d66024e-378f-4433-bc75-9088982a58a9" containerName="dnsmasq-dns" containerID="cri-o://b7f1944245f8d45a397ef5ccdcb3ad3132508559c41047d0a95422580ee8e502" gracePeriod=10 Dec 09 15:50:05 crc kubenswrapper[4894]: I1209 15:50:05.897520 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 09 15:50:05 crc kubenswrapper[4894]: I1209 15:50:05.954168 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.131775 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.148410 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-bbb4b9577-djjjb" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.249585 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-56fb46ccd4-rwfjk"] Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.250035 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-56fb46ccd4-rwfjk" podUID="03372157-f0dd-4db7-8662-3b110ff31815" containerName="neutron-api" containerID="cri-o://2e721c48a84006dcd7ba5afbee0b816cbf4fc8ec61bb683872a4cac2c7e906c2" gracePeriod=30 Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.250474 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-56fb46ccd4-rwfjk" podUID="03372157-f0dd-4db7-8662-3b110ff31815" containerName="neutron-httpd" containerID="cri-o://f1b7ca4463ac41d9e5ca80297e98942be3f547362d0beba09e8bd697366f6f4b" gracePeriod=30 Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.267911 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f69459849-bjgnn" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.284127 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.303681 4894 generic.go:334] "Generic (PLEG): container finished" podID="7d66024e-378f-4433-bc75-9088982a58a9" containerID="b7f1944245f8d45a397ef5ccdcb3ad3132508559c41047d0a95422580ee8e502" exitCode=0 Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.303764 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7f69459849-bjgnn" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.303787 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f69459849-bjgnn" event={"ID":"7d66024e-378f-4433-bc75-9088982a58a9","Type":"ContainerDied","Data":"b7f1944245f8d45a397ef5ccdcb3ad3132508559c41047d0a95422580ee8e502"} Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.303866 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7f69459849-bjgnn" event={"ID":"7d66024e-378f-4433-bc75-9088982a58a9","Type":"ContainerDied","Data":"b690b9ce22591250c9278450e01178bf9649fbde2f26bc63ec4a3c76e758e757"} Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.303888 4894 scope.go:117] "RemoveContainer" containerID="b7f1944245f8d45a397ef5ccdcb3ad3132508559c41047d0a95422580ee8e502" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.310850 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="f8c380b2-49de-4630-80e9-15c73cd99045" containerName="cinder-scheduler" containerID="cri-o://cd3e93d7b9ee1327ddfcd05f826c2a2212772a728777956cc2706bfa377b377d" gracePeriod=30 Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.311177 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="f8c380b2-49de-4630-80e9-15c73cd99045" containerName="probe" containerID="cri-o://b3e7d4a21768e3a06683a4ec83b56a18c03e4d8da51242eb72a21e484736b54d" gracePeriod=30 Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.311397 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"58e3314f-76b0-4e83-93ce-39c342c44acb","Type":"ContainerStarted","Data":"0ca60a3d87e35c9eab8eadbb0d8031370d42d591170f373c6552a9889047f261"} Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.311483 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.333136 4894 scope.go:117] "RemoveContainer" containerID="655772a72b1edb3848a5fafd5a82eb3b2bd6de7d6b6d50c6d38a0d76adc5f8c6" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.361969 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=3.361951481 podStartE2EDuration="3.361951481s" podCreationTimestamp="2025-12-09 15:50:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:50:06.351733064 +0000 UTC m=+1100.670943733" watchObservedRunningTime="2025-12-09 15:50:06.361951481 +0000 UTC m=+1100.681162150" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.375613 4894 scope.go:117] "RemoveContainer" containerID="b7f1944245f8d45a397ef5ccdcb3ad3132508559c41047d0a95422580ee8e502" Dec 09 15:50:06 crc kubenswrapper[4894]: E1209 15:50:06.378945 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7f1944245f8d45a397ef5ccdcb3ad3132508559c41047d0a95422580ee8e502\": container with ID starting with b7f1944245f8d45a397ef5ccdcb3ad3132508559c41047d0a95422580ee8e502 not found: ID does not exist" containerID="b7f1944245f8d45a397ef5ccdcb3ad3132508559c41047d0a95422580ee8e502" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.378983 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7f1944245f8d45a397ef5ccdcb3ad3132508559c41047d0a95422580ee8e502"} err="failed to get container status \"b7f1944245f8d45a397ef5ccdcb3ad3132508559c41047d0a95422580ee8e502\": rpc error: code = NotFound desc = could not find container \"b7f1944245f8d45a397ef5ccdcb3ad3132508559c41047d0a95422580ee8e502\": container with ID starting with b7f1944245f8d45a397ef5ccdcb3ad3132508559c41047d0a95422580ee8e502 not found: ID does not exist" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.379007 4894 scope.go:117] "RemoveContainer" containerID="655772a72b1edb3848a5fafd5a82eb3b2bd6de7d6b6d50c6d38a0d76adc5f8c6" Dec 09 15:50:06 crc kubenswrapper[4894]: E1209 15:50:06.379340 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"655772a72b1edb3848a5fafd5a82eb3b2bd6de7d6b6d50c6d38a0d76adc5f8c6\": container with ID starting with 655772a72b1edb3848a5fafd5a82eb3b2bd6de7d6b6d50c6d38a0d76adc5f8c6 not found: ID does not exist" containerID="655772a72b1edb3848a5fafd5a82eb3b2bd6de7d6b6d50c6d38a0d76adc5f8c6" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.379362 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"655772a72b1edb3848a5fafd5a82eb3b2bd6de7d6b6d50c6d38a0d76adc5f8c6"} err="failed to get container status \"655772a72b1edb3848a5fafd5a82eb3b2bd6de7d6b6d50c6d38a0d76adc5f8c6\": rpc error: code = NotFound desc = could not find container \"655772a72b1edb3848a5fafd5a82eb3b2bd6de7d6b6d50c6d38a0d76adc5f8c6\": container with ID starting with 655772a72b1edb3848a5fafd5a82eb3b2bd6de7d6b6d50c6d38a0d76adc5f8c6 not found: ID does not exist" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.413532 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-ovsdbserver-sb\") pod \"7d66024e-378f-4433-bc75-9088982a58a9\" (UID: \"7d66024e-378f-4433-bc75-9088982a58a9\") " Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.413614 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-config\") pod \"7d66024e-378f-4433-bc75-9088982a58a9\" (UID: \"7d66024e-378f-4433-bc75-9088982a58a9\") " Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.413708 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-dns-svc\") pod \"7d66024e-378f-4433-bc75-9088982a58a9\" (UID: \"7d66024e-378f-4433-bc75-9088982a58a9\") " Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.413768 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-ovsdbserver-nb\") pod \"7d66024e-378f-4433-bc75-9088982a58a9\" (UID: \"7d66024e-378f-4433-bc75-9088982a58a9\") " Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.413820 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tmqk6\" (UniqueName: \"kubernetes.io/projected/7d66024e-378f-4433-bc75-9088982a58a9-kube-api-access-tmqk6\") pod \"7d66024e-378f-4433-bc75-9088982a58a9\" (UID: \"7d66024e-378f-4433-bc75-9088982a58a9\") " Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.454041 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7d66024e-378f-4433-bc75-9088982a58a9-kube-api-access-tmqk6" (OuterVolumeSpecName: "kube-api-access-tmqk6") pod "7d66024e-378f-4433-bc75-9088982a58a9" (UID: "7d66024e-378f-4433-bc75-9088982a58a9"). InnerVolumeSpecName "kube-api-access-tmqk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.491753 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7d66024e-378f-4433-bc75-9088982a58a9" (UID: "7d66024e-378f-4433-bc75-9088982a58a9"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.493373 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7d66024e-378f-4433-bc75-9088982a58a9" (UID: "7d66024e-378f-4433-bc75-9088982a58a9"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.515938 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.515974 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tmqk6\" (UniqueName: \"kubernetes.io/projected/7d66024e-378f-4433-bc75-9088982a58a9-kube-api-access-tmqk6\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.515985 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.517292 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7d66024e-378f-4433-bc75-9088982a58a9" (UID: "7d66024e-378f-4433-bc75-9088982a58a9"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.521327 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-config" (OuterVolumeSpecName: "config") pod "7d66024e-378f-4433-bc75-9088982a58a9" (UID: "7d66024e-378f-4433-bc75-9088982a58a9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.617251 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.617290 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7d66024e-378f-4433-bc75-9088982a58a9-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.659040 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7f69459849-bjgnn"] Dec 09 15:50:06 crc kubenswrapper[4894]: I1209 15:50:06.697219 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7f69459849-bjgnn"] Dec 09 15:50:07 crc kubenswrapper[4894]: I1209 15:50:07.323211 4894 generic.go:334] "Generic (PLEG): container finished" podID="03372157-f0dd-4db7-8662-3b110ff31815" containerID="f1b7ca4463ac41d9e5ca80297e98942be3f547362d0beba09e8bd697366f6f4b" exitCode=0 Dec 09 15:50:07 crc kubenswrapper[4894]: I1209 15:50:07.323465 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56fb46ccd4-rwfjk" event={"ID":"03372157-f0dd-4db7-8662-3b110ff31815","Type":"ContainerDied","Data":"f1b7ca4463ac41d9e5ca80297e98942be3f547362d0beba09e8bd697366f6f4b"} Dec 09 15:50:07 crc kubenswrapper[4894]: I1209 15:50:07.326614 4894 generic.go:334] "Generic (PLEG): container finished" podID="0e41ac7e-15c8-4a9e-ae96-2e68b055ea16" containerID="1b3e955a174036764325c570292b5dde1384ef1df60f2868c564915e82f08177" exitCode=0 Dec 09 15:50:07 crc kubenswrapper[4894]: I1209 15:50:07.326664 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-56f65bfbb-tv5ld" event={"ID":"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16","Type":"ContainerDied","Data":"1b3e955a174036764325c570292b5dde1384ef1df60f2868c564915e82f08177"} Dec 09 15:50:07 crc kubenswrapper[4894]: I1209 15:50:07.329676 4894 generic.go:334] "Generic (PLEG): container finished" podID="f8c380b2-49de-4630-80e9-15c73cd99045" containerID="b3e7d4a21768e3a06683a4ec83b56a18c03e4d8da51242eb72a21e484736b54d" exitCode=0 Dec 09 15:50:07 crc kubenswrapper[4894]: I1209 15:50:07.329761 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f8c380b2-49de-4630-80e9-15c73cd99045","Type":"ContainerDied","Data":"b3e7d4a21768e3a06683a4ec83b56a18c03e4d8da51242eb72a21e484736b54d"} Dec 09 15:50:08 crc kubenswrapper[4894]: I1209 15:50:08.123032 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d66024e-378f-4433-bc75-9088982a58a9" path="/var/lib/kubelet/pods/7d66024e-378f-4433-bc75-9088982a58a9/volumes" Dec 09 15:50:08 crc kubenswrapper[4894]: I1209 15:50:08.564425 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-56f65bfbb-tv5ld" podUID="0e41ac7e-15c8-4a9e-ae96-2e68b055ea16" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.128780 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.196895 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-scripts\") pod \"f8c380b2-49de-4630-80e9-15c73cd99045\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.197215 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-config-data-custom\") pod \"f8c380b2-49de-4630-80e9-15c73cd99045\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.197260 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-config-data\") pod \"f8c380b2-49de-4630-80e9-15c73cd99045\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.197294 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8c380b2-49de-4630-80e9-15c73cd99045-etc-machine-id\") pod \"f8c380b2-49de-4630-80e9-15c73cd99045\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.197337 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqp46\" (UniqueName: \"kubernetes.io/projected/f8c380b2-49de-4630-80e9-15c73cd99045-kube-api-access-qqp46\") pod \"f8c380b2-49de-4630-80e9-15c73cd99045\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.197372 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-combined-ca-bundle\") pod \"f8c380b2-49de-4630-80e9-15c73cd99045\" (UID: \"f8c380b2-49de-4630-80e9-15c73cd99045\") " Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.204399 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f8c380b2-49de-4630-80e9-15c73cd99045-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "f8c380b2-49de-4630-80e9-15c73cd99045" (UID: "f8c380b2-49de-4630-80e9-15c73cd99045"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.205747 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-scripts" (OuterVolumeSpecName: "scripts") pod "f8c380b2-49de-4630-80e9-15c73cd99045" (UID: "f8c380b2-49de-4630-80e9-15c73cd99045"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.206770 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8c380b2-49de-4630-80e9-15c73cd99045-kube-api-access-qqp46" (OuterVolumeSpecName: "kube-api-access-qqp46") pod "f8c380b2-49de-4630-80e9-15c73cd99045" (UID: "f8c380b2-49de-4630-80e9-15c73cd99045"). InnerVolumeSpecName "kube-api-access-qqp46". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.222571 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "f8c380b2-49de-4630-80e9-15c73cd99045" (UID: "f8c380b2-49de-4630-80e9-15c73cd99045"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.260661 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f8c380b2-49de-4630-80e9-15c73cd99045" (UID: "f8c380b2-49de-4630-80e9-15c73cd99045"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.299037 4894 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/f8c380b2-49de-4630-80e9-15c73cd99045-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.299073 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqp46\" (UniqueName: \"kubernetes.io/projected/f8c380b2-49de-4630-80e9-15c73cd99045-kube-api-access-qqp46\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.299085 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.299094 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.299103 4894 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.314752 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-config-data" (OuterVolumeSpecName: "config-data") pod "f8c380b2-49de-4630-80e9-15c73cd99045" (UID: "f8c380b2-49de-4630-80e9-15c73cd99045"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.362574 4894 generic.go:334] "Generic (PLEG): container finished" podID="f8c380b2-49de-4630-80e9-15c73cd99045" containerID="cd3e93d7b9ee1327ddfcd05f826c2a2212772a728777956cc2706bfa377b377d" exitCode=0 Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.362649 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f8c380b2-49de-4630-80e9-15c73cd99045","Type":"ContainerDied","Data":"cd3e93d7b9ee1327ddfcd05f826c2a2212772a728777956cc2706bfa377b377d"} Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.362683 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"f8c380b2-49de-4630-80e9-15c73cd99045","Type":"ContainerDied","Data":"f8aeec9de6975d944e48cedf195b0b7cb434ee40bcda5accecffffb3d8f4bd59"} Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.362703 4894 scope.go:117] "RemoveContainer" containerID="b3e7d4a21768e3a06683a4ec83b56a18c03e4d8da51242eb72a21e484736b54d" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.362880 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.403200 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8c380b2-49de-4630-80e9-15c73cd99045-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.406687 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.414448 4894 scope.go:117] "RemoveContainer" containerID="cd3e93d7b9ee1327ddfcd05f826c2a2212772a728777956cc2706bfa377b377d" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.439992 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.445745 4894 scope.go:117] "RemoveContainer" containerID="b3e7d4a21768e3a06683a4ec83b56a18c03e4d8da51242eb72a21e484736b54d" Dec 09 15:50:11 crc kubenswrapper[4894]: E1209 15:50:11.454849 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b3e7d4a21768e3a06683a4ec83b56a18c03e4d8da51242eb72a21e484736b54d\": container with ID starting with b3e7d4a21768e3a06683a4ec83b56a18c03e4d8da51242eb72a21e484736b54d not found: ID does not exist" containerID="b3e7d4a21768e3a06683a4ec83b56a18c03e4d8da51242eb72a21e484736b54d" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.454917 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b3e7d4a21768e3a06683a4ec83b56a18c03e4d8da51242eb72a21e484736b54d"} err="failed to get container status \"b3e7d4a21768e3a06683a4ec83b56a18c03e4d8da51242eb72a21e484736b54d\": rpc error: code = NotFound desc = could not find container \"b3e7d4a21768e3a06683a4ec83b56a18c03e4d8da51242eb72a21e484736b54d\": container with ID starting with b3e7d4a21768e3a06683a4ec83b56a18c03e4d8da51242eb72a21e484736b54d not found: ID does not exist" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.454955 4894 scope.go:117] "RemoveContainer" containerID="cd3e93d7b9ee1327ddfcd05f826c2a2212772a728777956cc2706bfa377b377d" Dec 09 15:50:11 crc kubenswrapper[4894]: E1209 15:50:11.458723 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cd3e93d7b9ee1327ddfcd05f826c2a2212772a728777956cc2706bfa377b377d\": container with ID starting with cd3e93d7b9ee1327ddfcd05f826c2a2212772a728777956cc2706bfa377b377d not found: ID does not exist" containerID="cd3e93d7b9ee1327ddfcd05f826c2a2212772a728777956cc2706bfa377b377d" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.458753 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cd3e93d7b9ee1327ddfcd05f826c2a2212772a728777956cc2706bfa377b377d"} err="failed to get container status \"cd3e93d7b9ee1327ddfcd05f826c2a2212772a728777956cc2706bfa377b377d\": rpc error: code = NotFound desc = could not find container \"cd3e93d7b9ee1327ddfcd05f826c2a2212772a728777956cc2706bfa377b377d\": container with ID starting with cd3e93d7b9ee1327ddfcd05f826c2a2212772a728777956cc2706bfa377b377d not found: ID does not exist" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.467294 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 15:50:11 crc kubenswrapper[4894]: E1209 15:50:11.467744 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d66024e-378f-4433-bc75-9088982a58a9" containerName="init" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.467770 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d66024e-378f-4433-bc75-9088982a58a9" containerName="init" Dec 09 15:50:11 crc kubenswrapper[4894]: E1209 15:50:11.467809 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7d66024e-378f-4433-bc75-9088982a58a9" containerName="dnsmasq-dns" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.467819 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7d66024e-378f-4433-bc75-9088982a58a9" containerName="dnsmasq-dns" Dec 09 15:50:11 crc kubenswrapper[4894]: E1209 15:50:11.467834 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c380b2-49de-4630-80e9-15c73cd99045" containerName="cinder-scheduler" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.467842 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c380b2-49de-4630-80e9-15c73cd99045" containerName="cinder-scheduler" Dec 09 15:50:11 crc kubenswrapper[4894]: E1209 15:50:11.467863 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8c380b2-49de-4630-80e9-15c73cd99045" containerName="probe" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.467870 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8c380b2-49de-4630-80e9-15c73cd99045" containerName="probe" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.468071 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8c380b2-49de-4630-80e9-15c73cd99045" containerName="probe" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.468097 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8c380b2-49de-4630-80e9-15c73cd99045" containerName="cinder-scheduler" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.468115 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7d66024e-378f-4433-bc75-9088982a58a9" containerName="dnsmasq-dns" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.469278 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.475081 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.475841 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.609122 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fe06c113-787a-4daa-bba6-2c9ca0b1b4d3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fe06c113-787a-4daa-bba6-2c9ca0b1b4d3\") " pod="openstack/cinder-scheduler-0" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.609190 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe06c113-787a-4daa-bba6-2c9ca0b1b4d3-config-data\") pod \"cinder-scheduler-0\" (UID: \"fe06c113-787a-4daa-bba6-2c9ca0b1b4d3\") " pod="openstack/cinder-scheduler-0" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.609237 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe06c113-787a-4daa-bba6-2c9ca0b1b4d3-scripts\") pod \"cinder-scheduler-0\" (UID: \"fe06c113-787a-4daa-bba6-2c9ca0b1b4d3\") " pod="openstack/cinder-scheduler-0" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.609276 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe06c113-787a-4daa-bba6-2c9ca0b1b4d3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fe06c113-787a-4daa-bba6-2c9ca0b1b4d3\") " pod="openstack/cinder-scheduler-0" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.609373 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe06c113-787a-4daa-bba6-2c9ca0b1b4d3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fe06c113-787a-4daa-bba6-2c9ca0b1b4d3\") " pod="openstack/cinder-scheduler-0" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.609480 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqxz8\" (UniqueName: \"kubernetes.io/projected/fe06c113-787a-4daa-bba6-2c9ca0b1b4d3-kube-api-access-sqxz8\") pod \"cinder-scheduler-0\" (UID: \"fe06c113-787a-4daa-bba6-2c9ca0b1b4d3\") " pod="openstack/cinder-scheduler-0" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.711722 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fe06c113-787a-4daa-bba6-2c9ca0b1b4d3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fe06c113-787a-4daa-bba6-2c9ca0b1b4d3\") " pod="openstack/cinder-scheduler-0" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.711787 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe06c113-787a-4daa-bba6-2c9ca0b1b4d3-config-data\") pod \"cinder-scheduler-0\" (UID: \"fe06c113-787a-4daa-bba6-2c9ca0b1b4d3\") " pod="openstack/cinder-scheduler-0" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.711837 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe06c113-787a-4daa-bba6-2c9ca0b1b4d3-scripts\") pod \"cinder-scheduler-0\" (UID: \"fe06c113-787a-4daa-bba6-2c9ca0b1b4d3\") " pod="openstack/cinder-scheduler-0" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.711876 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe06c113-787a-4daa-bba6-2c9ca0b1b4d3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fe06c113-787a-4daa-bba6-2c9ca0b1b4d3\") " pod="openstack/cinder-scheduler-0" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.711877 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/fe06c113-787a-4daa-bba6-2c9ca0b1b4d3-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"fe06c113-787a-4daa-bba6-2c9ca0b1b4d3\") " pod="openstack/cinder-scheduler-0" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.712020 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe06c113-787a-4daa-bba6-2c9ca0b1b4d3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fe06c113-787a-4daa-bba6-2c9ca0b1b4d3\") " pod="openstack/cinder-scheduler-0" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.712234 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqxz8\" (UniqueName: \"kubernetes.io/projected/fe06c113-787a-4daa-bba6-2c9ca0b1b4d3-kube-api-access-sqxz8\") pod \"cinder-scheduler-0\" (UID: \"fe06c113-787a-4daa-bba6-2c9ca0b1b4d3\") " pod="openstack/cinder-scheduler-0" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.716507 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/fe06c113-787a-4daa-bba6-2c9ca0b1b4d3-scripts\") pod \"cinder-scheduler-0\" (UID: \"fe06c113-787a-4daa-bba6-2c9ca0b1b4d3\") " pod="openstack/cinder-scheduler-0" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.716536 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe06c113-787a-4daa-bba6-2c9ca0b1b4d3-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"fe06c113-787a-4daa-bba6-2c9ca0b1b4d3\") " pod="openstack/cinder-scheduler-0" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.716872 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe06c113-787a-4daa-bba6-2c9ca0b1b4d3-config-data\") pod \"cinder-scheduler-0\" (UID: \"fe06c113-787a-4daa-bba6-2c9ca0b1b4d3\") " pod="openstack/cinder-scheduler-0" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.718213 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fe06c113-787a-4daa-bba6-2c9ca0b1b4d3-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"fe06c113-787a-4daa-bba6-2c9ca0b1b4d3\") " pod="openstack/cinder-scheduler-0" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.739356 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqxz8\" (UniqueName: \"kubernetes.io/projected/fe06c113-787a-4daa-bba6-2c9ca0b1b4d3-kube-api-access-sqxz8\") pod \"cinder-scheduler-0\" (UID: \"fe06c113-787a-4daa-bba6-2c9ca0b1b4d3\") " pod="openstack/cinder-scheduler-0" Dec 09 15:50:11 crc kubenswrapper[4894]: I1209 15:50:11.789242 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Dec 09 15:50:12 crc kubenswrapper[4894]: I1209 15:50:12.119943 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8c380b2-49de-4630-80e9-15c73cd99045" path="/var/lib/kubelet/pods/f8c380b2-49de-4630-80e9-15c73cd99045/volumes" Dec 09 15:50:12 crc kubenswrapper[4894]: I1209 15:50:12.257260 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:50:12 crc kubenswrapper[4894]: I1209 15:50:12.257318 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:50:12 crc kubenswrapper[4894]: I1209 15:50:12.306467 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Dec 09 15:50:12 crc kubenswrapper[4894]: I1209 15:50:12.372264 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fe06c113-787a-4daa-bba6-2c9ca0b1b4d3","Type":"ContainerStarted","Data":"5c1c19faf547b4477139a480a3a518fffb4b873ea9f998a4c51e31b20a10321e"} Dec 09 15:50:12 crc kubenswrapper[4894]: I1209 15:50:12.662007 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:50:12 crc kubenswrapper[4894]: I1209 15:50:12.663254 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-59f4fb4758-w8ndv" Dec 09 15:50:13 crc kubenswrapper[4894]: I1209 15:50:13.040385 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:13 crc kubenswrapper[4894]: I1209 15:50:13.188529 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-5df7cc9c4d-kfjtd" Dec 09 15:50:13 crc kubenswrapper[4894]: I1209 15:50:13.287935 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-569f9c8ccb-pkfj6"] Dec 09 15:50:13 crc kubenswrapper[4894]: I1209 15:50:13.288162 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-569f9c8ccb-pkfj6" podUID="5ecec7eb-d196-471a-9e5c-4ae0ce5ee992" containerName="barbican-api-log" containerID="cri-o://11ee86bb97843f66a3b73c31e5f991932bfb9b69347db6f215e2c4493afc94fe" gracePeriod=30 Dec 09 15:50:13 crc kubenswrapper[4894]: I1209 15:50:13.288284 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-569f9c8ccb-pkfj6" podUID="5ecec7eb-d196-471a-9e5c-4ae0ce5ee992" containerName="barbican-api" containerID="cri-o://ed959f038ef472ffa42f39eb22c506846d5127725413646ef9b8a0ecbf5c25be" gracePeriod=30 Dec 09 15:50:13 crc kubenswrapper[4894]: I1209 15:50:13.387030 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fe06c113-787a-4daa-bba6-2c9ca0b1b4d3","Type":"ContainerStarted","Data":"c201ac257104789f8565c088ed0db48d75773e5ade34108d8fb2d45f22794f35"} Dec 09 15:50:14 crc kubenswrapper[4894]: I1209 15:50:14.396158 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"fe06c113-787a-4daa-bba6-2c9ca0b1b4d3","Type":"ContainerStarted","Data":"3262e1f3ec0b626eb7091476353e1c95732ee3fa5dfc1608ecbff3dda4cb3352"} Dec 09 15:50:14 crc kubenswrapper[4894]: I1209 15:50:14.399587 4894 generic.go:334] "Generic (PLEG): container finished" podID="5ecec7eb-d196-471a-9e5c-4ae0ce5ee992" containerID="11ee86bb97843f66a3b73c31e5f991932bfb9b69347db6f215e2c4493afc94fe" exitCode=143 Dec 09 15:50:14 crc kubenswrapper[4894]: I1209 15:50:14.399653 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-569f9c8ccb-pkfj6" event={"ID":"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992","Type":"ContainerDied","Data":"11ee86bb97843f66a3b73c31e5f991932bfb9b69347db6f215e2c4493afc94fe"} Dec 09 15:50:14 crc kubenswrapper[4894]: I1209 15:50:14.468057 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=3.468035836 podStartE2EDuration="3.468035836s" podCreationTimestamp="2025-12-09 15:50:11 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:50:14.459417729 +0000 UTC m=+1108.778628398" watchObservedRunningTime="2025-12-09 15:50:14.468035836 +0000 UTC m=+1108.787246515" Dec 09 15:50:15 crc kubenswrapper[4894]: I1209 15:50:15.598244 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-ccc8cf6fb-9hvcd" Dec 09 15:50:16 crc kubenswrapper[4894]: I1209 15:50:16.266002 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Dec 09 15:50:16 crc kubenswrapper[4894]: I1209 15:50:16.479175 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-569f9c8ccb-pkfj6" podUID="5ecec7eb-d196-471a-9e5c-4ae0ce5ee992" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.154:9311/healthcheck\": read tcp 10.217.0.2:46084->10.217.0.154:9311: read: connection reset by peer" Dec 09 15:50:16 crc kubenswrapper[4894]: I1209 15:50:16.479200 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-569f9c8ccb-pkfj6" podUID="5ecec7eb-d196-471a-9e5c-4ae0ce5ee992" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.154:9311/healthcheck\": read tcp 10.217.0.2:46080->10.217.0.154:9311: read: connection reset by peer" Dec 09 15:50:16 crc kubenswrapper[4894]: I1209 15:50:16.790212 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.357009 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.425474 4894 generic.go:334] "Generic (PLEG): container finished" podID="5ecec7eb-d196-471a-9e5c-4ae0ce5ee992" containerID="ed959f038ef472ffa42f39eb22c506846d5127725413646ef9b8a0ecbf5c25be" exitCode=0 Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.425526 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-569f9c8ccb-pkfj6" event={"ID":"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992","Type":"ContainerDied","Data":"ed959f038ef472ffa42f39eb22c506846d5127725413646ef9b8a0ecbf5c25be"} Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.425568 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-569f9c8ccb-pkfj6" Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.425596 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-569f9c8ccb-pkfj6" event={"ID":"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992","Type":"ContainerDied","Data":"4d4aed6ba4c61e4e05b19fe908160be8df98cfde903dbb0f43021286e99abe42"} Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.425617 4894 scope.go:117] "RemoveContainer" containerID="ed959f038ef472ffa42f39eb22c506846d5127725413646ef9b8a0ecbf5c25be" Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.453010 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sr87v\" (UniqueName: \"kubernetes.io/projected/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-kube-api-access-sr87v\") pod \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\" (UID: \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\") " Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.453108 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-config-data\") pod \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\" (UID: \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\") " Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.453172 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-logs\") pod \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\" (UID: \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\") " Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.453203 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-config-data-custom\") pod \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\" (UID: \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\") " Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.453264 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-combined-ca-bundle\") pod \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\" (UID: \"5ecec7eb-d196-471a-9e5c-4ae0ce5ee992\") " Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.453925 4894 scope.go:117] "RemoveContainer" containerID="11ee86bb97843f66a3b73c31e5f991932bfb9b69347db6f215e2c4493afc94fe" Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.455707 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-logs" (OuterVolumeSpecName: "logs") pod "5ecec7eb-d196-471a-9e5c-4ae0ce5ee992" (UID: "5ecec7eb-d196-471a-9e5c-4ae0ce5ee992"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.460601 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5ecec7eb-d196-471a-9e5c-4ae0ce5ee992" (UID: "5ecec7eb-d196-471a-9e5c-4ae0ce5ee992"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.470260 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-kube-api-access-sr87v" (OuterVolumeSpecName: "kube-api-access-sr87v") pod "5ecec7eb-d196-471a-9e5c-4ae0ce5ee992" (UID: "5ecec7eb-d196-471a-9e5c-4ae0ce5ee992"). InnerVolumeSpecName "kube-api-access-sr87v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.485402 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5ecec7eb-d196-471a-9e5c-4ae0ce5ee992" (UID: "5ecec7eb-d196-471a-9e5c-4ae0ce5ee992"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.511967 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-config-data" (OuterVolumeSpecName: "config-data") pod "5ecec7eb-d196-471a-9e5c-4ae0ce5ee992" (UID: "5ecec7eb-d196-471a-9e5c-4ae0ce5ee992"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.556235 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sr87v\" (UniqueName: \"kubernetes.io/projected/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-kube-api-access-sr87v\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.556276 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.556313 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-logs\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.556325 4894 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.556337 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.570089 4894 scope.go:117] "RemoveContainer" containerID="ed959f038ef472ffa42f39eb22c506846d5127725413646ef9b8a0ecbf5c25be" Dec 09 15:50:17 crc kubenswrapper[4894]: E1209 15:50:17.570605 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed959f038ef472ffa42f39eb22c506846d5127725413646ef9b8a0ecbf5c25be\": container with ID starting with ed959f038ef472ffa42f39eb22c506846d5127725413646ef9b8a0ecbf5c25be not found: ID does not exist" containerID="ed959f038ef472ffa42f39eb22c506846d5127725413646ef9b8a0ecbf5c25be" Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.570758 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed959f038ef472ffa42f39eb22c506846d5127725413646ef9b8a0ecbf5c25be"} err="failed to get container status \"ed959f038ef472ffa42f39eb22c506846d5127725413646ef9b8a0ecbf5c25be\": rpc error: code = NotFound desc = could not find container \"ed959f038ef472ffa42f39eb22c506846d5127725413646ef9b8a0ecbf5c25be\": container with ID starting with ed959f038ef472ffa42f39eb22c506846d5127725413646ef9b8a0ecbf5c25be not found: ID does not exist" Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.570790 4894 scope.go:117] "RemoveContainer" containerID="11ee86bb97843f66a3b73c31e5f991932bfb9b69347db6f215e2c4493afc94fe" Dec 09 15:50:17 crc kubenswrapper[4894]: E1209 15:50:17.571162 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"11ee86bb97843f66a3b73c31e5f991932bfb9b69347db6f215e2c4493afc94fe\": container with ID starting with 11ee86bb97843f66a3b73c31e5f991932bfb9b69347db6f215e2c4493afc94fe not found: ID does not exist" containerID="11ee86bb97843f66a3b73c31e5f991932bfb9b69347db6f215e2c4493afc94fe" Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.571194 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"11ee86bb97843f66a3b73c31e5f991932bfb9b69347db6f215e2c4493afc94fe"} err="failed to get container status \"11ee86bb97843f66a3b73c31e5f991932bfb9b69347db6f215e2c4493afc94fe\": rpc error: code = NotFound desc = could not find container \"11ee86bb97843f66a3b73c31e5f991932bfb9b69347db6f215e2c4493afc94fe\": container with ID starting with 11ee86bb97843f66a3b73c31e5f991932bfb9b69347db6f215e2c4493afc94fe not found: ID does not exist" Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.759940 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-569f9c8ccb-pkfj6"] Dec 09 15:50:17 crc kubenswrapper[4894]: I1209 15:50:17.778803 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-569f9c8ccb-pkfj6"] Dec 09 15:50:18 crc kubenswrapper[4894]: I1209 15:50:18.121479 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ecec7eb-d196-471a-9e5c-4ae0ce5ee992" path="/var/lib/kubelet/pods/5ecec7eb-d196-471a-9e5c-4ae0ce5ee992/volumes" Dec 09 15:50:18 crc kubenswrapper[4894]: I1209 15:50:18.443097 4894 generic.go:334] "Generic (PLEG): container finished" podID="03372157-f0dd-4db7-8662-3b110ff31815" containerID="2e721c48a84006dcd7ba5afbee0b816cbf4fc8ec61bb683872a4cac2c7e906c2" exitCode=0 Dec 09 15:50:18 crc kubenswrapper[4894]: I1209 15:50:18.443249 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56fb46ccd4-rwfjk" event={"ID":"03372157-f0dd-4db7-8662-3b110ff31815","Type":"ContainerDied","Data":"2e721c48a84006dcd7ba5afbee0b816cbf4fc8ec61bb683872a4cac2c7e906c2"} Dec 09 15:50:18 crc kubenswrapper[4894]: I1209 15:50:18.564320 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-56f65bfbb-tv5ld" podUID="0e41ac7e-15c8-4a9e-ae96-2e68b055ea16" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 09 15:50:18 crc kubenswrapper[4894]: I1209 15:50:18.687226 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56fb46ccd4-rwfjk" Dec 09 15:50:18 crc kubenswrapper[4894]: I1209 15:50:18.779299 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kfm2\" (UniqueName: \"kubernetes.io/projected/03372157-f0dd-4db7-8662-3b110ff31815-kube-api-access-6kfm2\") pod \"03372157-f0dd-4db7-8662-3b110ff31815\" (UID: \"03372157-f0dd-4db7-8662-3b110ff31815\") " Dec 09 15:50:18 crc kubenswrapper[4894]: I1209 15:50:18.779543 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-config\") pod \"03372157-f0dd-4db7-8662-3b110ff31815\" (UID: \"03372157-f0dd-4db7-8662-3b110ff31815\") " Dec 09 15:50:18 crc kubenswrapper[4894]: I1209 15:50:18.779583 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-combined-ca-bundle\") pod \"03372157-f0dd-4db7-8662-3b110ff31815\" (UID: \"03372157-f0dd-4db7-8662-3b110ff31815\") " Dec 09 15:50:18 crc kubenswrapper[4894]: I1209 15:50:18.779633 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-ovndb-tls-certs\") pod \"03372157-f0dd-4db7-8662-3b110ff31815\" (UID: \"03372157-f0dd-4db7-8662-3b110ff31815\") " Dec 09 15:50:18 crc kubenswrapper[4894]: I1209 15:50:18.779702 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-httpd-config\") pod \"03372157-f0dd-4db7-8662-3b110ff31815\" (UID: \"03372157-f0dd-4db7-8662-3b110ff31815\") " Dec 09 15:50:18 crc kubenswrapper[4894]: I1209 15:50:18.787780 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "03372157-f0dd-4db7-8662-3b110ff31815" (UID: "03372157-f0dd-4db7-8662-3b110ff31815"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:18 crc kubenswrapper[4894]: I1209 15:50:18.791842 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/03372157-f0dd-4db7-8662-3b110ff31815-kube-api-access-6kfm2" (OuterVolumeSpecName: "kube-api-access-6kfm2") pod "03372157-f0dd-4db7-8662-3b110ff31815" (UID: "03372157-f0dd-4db7-8662-3b110ff31815"). InnerVolumeSpecName "kube-api-access-6kfm2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:50:18 crc kubenswrapper[4894]: I1209 15:50:18.861353 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-config" (OuterVolumeSpecName: "config") pod "03372157-f0dd-4db7-8662-3b110ff31815" (UID: "03372157-f0dd-4db7-8662-3b110ff31815"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:18 crc kubenswrapper[4894]: I1209 15:50:18.867145 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "03372157-f0dd-4db7-8662-3b110ff31815" (UID: "03372157-f0dd-4db7-8662-3b110ff31815"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:18 crc kubenswrapper[4894]: I1209 15:50:18.867883 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "03372157-f0dd-4db7-8662-3b110ff31815" (UID: "03372157-f0dd-4db7-8662-3b110ff31815"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:18 crc kubenswrapper[4894]: I1209 15:50:18.882070 4894 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:18 crc kubenswrapper[4894]: I1209 15:50:18.882104 4894 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-httpd-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:18 crc kubenswrapper[4894]: I1209 15:50:18.882114 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kfm2\" (UniqueName: \"kubernetes.io/projected/03372157-f0dd-4db7-8662-3b110ff31815-kube-api-access-6kfm2\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:18 crc kubenswrapper[4894]: I1209 15:50:18.882126 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:18 crc kubenswrapper[4894]: I1209 15:50:18.882136 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/03372157-f0dd-4db7-8662-3b110ff31815-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.454879 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-56fb46ccd4-rwfjk" event={"ID":"03372157-f0dd-4db7-8662-3b110ff31815","Type":"ContainerDied","Data":"335e3809b8720ed0c732689163dbea75739835ceae0bae80a73f7acb7c2492d9"} Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.454927 4894 scope.go:117] "RemoveContainer" containerID="f1b7ca4463ac41d9e5ca80297e98942be3f547362d0beba09e8bd697366f6f4b" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.454961 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-56fb46ccd4-rwfjk" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.483051 4894 scope.go:117] "RemoveContainer" containerID="2e721c48a84006dcd7ba5afbee0b816cbf4fc8ec61bb683872a4cac2c7e906c2" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.489821 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-56fb46ccd4-rwfjk"] Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.497403 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-56fb46ccd4-rwfjk"] Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.657118 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Dec 09 15:50:19 crc kubenswrapper[4894]: E1209 15:50:19.657578 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ecec7eb-d196-471a-9e5c-4ae0ce5ee992" containerName="barbican-api-log" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.657600 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ecec7eb-d196-471a-9e5c-4ae0ce5ee992" containerName="barbican-api-log" Dec 09 15:50:19 crc kubenswrapper[4894]: E1209 15:50:19.657622 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ecec7eb-d196-471a-9e5c-4ae0ce5ee992" containerName="barbican-api" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.657630 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ecec7eb-d196-471a-9e5c-4ae0ce5ee992" containerName="barbican-api" Dec 09 15:50:19 crc kubenswrapper[4894]: E1209 15:50:19.657719 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03372157-f0dd-4db7-8662-3b110ff31815" containerName="neutron-api" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.657728 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="03372157-f0dd-4db7-8662-3b110ff31815" containerName="neutron-api" Dec 09 15:50:19 crc kubenswrapper[4894]: E1209 15:50:19.657749 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="03372157-f0dd-4db7-8662-3b110ff31815" containerName="neutron-httpd" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.657759 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="03372157-f0dd-4db7-8662-3b110ff31815" containerName="neutron-httpd" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.657967 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="03372157-f0dd-4db7-8662-3b110ff31815" containerName="neutron-httpd" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.657990 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ecec7eb-d196-471a-9e5c-4ae0ce5ee992" containerName="barbican-api-log" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.658000 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="03372157-f0dd-4db7-8662-3b110ff31815" containerName="neutron-api" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.658015 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ecec7eb-d196-471a-9e5c-4ae0ce5ee992" containerName="barbican-api" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.658630 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.661552 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.661919 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.662843 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-8gvqs" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.665802 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.798384 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3752bfef-7b4f-4f5d-a449-a8b0819f6434-openstack-config-secret\") pod \"openstackclient\" (UID: \"3752bfef-7b4f-4f5d-a449-a8b0819f6434\") " pod="openstack/openstackclient" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.798458 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3752bfef-7b4f-4f5d-a449-a8b0819f6434-openstack-config\") pod \"openstackclient\" (UID: \"3752bfef-7b4f-4f5d-a449-a8b0819f6434\") " pod="openstack/openstackclient" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.798507 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjvb4\" (UniqueName: \"kubernetes.io/projected/3752bfef-7b4f-4f5d-a449-a8b0819f6434-kube-api-access-wjvb4\") pod \"openstackclient\" (UID: \"3752bfef-7b4f-4f5d-a449-a8b0819f6434\") " pod="openstack/openstackclient" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.798737 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3752bfef-7b4f-4f5d-a449-a8b0819f6434-combined-ca-bundle\") pod \"openstackclient\" (UID: \"3752bfef-7b4f-4f5d-a449-a8b0819f6434\") " pod="openstack/openstackclient" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.900951 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3752bfef-7b4f-4f5d-a449-a8b0819f6434-openstack-config\") pod \"openstackclient\" (UID: \"3752bfef-7b4f-4f5d-a449-a8b0819f6434\") " pod="openstack/openstackclient" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.901020 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjvb4\" (UniqueName: \"kubernetes.io/projected/3752bfef-7b4f-4f5d-a449-a8b0819f6434-kube-api-access-wjvb4\") pod \"openstackclient\" (UID: \"3752bfef-7b4f-4f5d-a449-a8b0819f6434\") " pod="openstack/openstackclient" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.901086 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3752bfef-7b4f-4f5d-a449-a8b0819f6434-combined-ca-bundle\") pod \"openstackclient\" (UID: \"3752bfef-7b4f-4f5d-a449-a8b0819f6434\") " pod="openstack/openstackclient" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.901161 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3752bfef-7b4f-4f5d-a449-a8b0819f6434-openstack-config-secret\") pod \"openstackclient\" (UID: \"3752bfef-7b4f-4f5d-a449-a8b0819f6434\") " pod="openstack/openstackclient" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.902014 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/3752bfef-7b4f-4f5d-a449-a8b0819f6434-openstack-config\") pod \"openstackclient\" (UID: \"3752bfef-7b4f-4f5d-a449-a8b0819f6434\") " pod="openstack/openstackclient" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.904876 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3752bfef-7b4f-4f5d-a449-a8b0819f6434-combined-ca-bundle\") pod \"openstackclient\" (UID: \"3752bfef-7b4f-4f5d-a449-a8b0819f6434\") " pod="openstack/openstackclient" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.906393 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3752bfef-7b4f-4f5d-a449-a8b0819f6434-openstack-config-secret\") pod \"openstackclient\" (UID: \"3752bfef-7b4f-4f5d-a449-a8b0819f6434\") " pod="openstack/openstackclient" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.922090 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjvb4\" (UniqueName: \"kubernetes.io/projected/3752bfef-7b4f-4f5d-a449-a8b0819f6434-kube-api-access-wjvb4\") pod \"openstackclient\" (UID: \"3752bfef-7b4f-4f5d-a449-a8b0819f6434\") " pod="openstack/openstackclient" Dec 09 15:50:19 crc kubenswrapper[4894]: I1209 15:50:19.989668 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Dec 09 15:50:20 crc kubenswrapper[4894]: I1209 15:50:20.117249 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="03372157-f0dd-4db7-8662-3b110ff31815" path="/var/lib/kubelet/pods/03372157-f0dd-4db7-8662-3b110ff31815/volumes" Dec 09 15:50:20 crc kubenswrapper[4894]: I1209 15:50:20.629608 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Dec 09 15:50:20 crc kubenswrapper[4894]: W1209 15:50:20.637163 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3752bfef_7b4f_4f5d_a449_a8b0819f6434.slice/crio-f0b8f544e3430a82971b1629d1b45e50c462992defed9e14e8a3ae87b827c9e4 WatchSource:0}: Error finding container f0b8f544e3430a82971b1629d1b45e50c462992defed9e14e8a3ae87b827c9e4: Status 404 returned error can't find the container with id f0b8f544e3430a82971b1629d1b45e50c462992defed9e14e8a3ae87b827c9e4 Dec 09 15:50:21 crc kubenswrapper[4894]: I1209 15:50:21.476439 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"3752bfef-7b4f-4f5d-a449-a8b0819f6434","Type":"ContainerStarted","Data":"f0b8f544e3430a82971b1629d1b45e50c462992defed9e14e8a3ae87b827c9e4"} Dec 09 15:50:22 crc kubenswrapper[4894]: I1209 15:50:22.084885 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Dec 09 15:50:26 crc kubenswrapper[4894]: I1209 15:50:26.603751 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 09 15:50:28 crc kubenswrapper[4894]: I1209 15:50:28.565265 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/horizon-56f65bfbb-tv5ld" podUID="0e41ac7e-15c8-4a9e-ae96-2e68b055ea16" containerName="horizon" probeResult="failure" output="Get \"https://10.217.0.143:8443/dashboard/auth/login/?next=/dashboard/\": dial tcp 10.217.0.143:8443: connect: connection refused" Dec 09 15:50:28 crc kubenswrapper[4894]: I1209 15:50:28.566835 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.195100 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.195959 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e1922e1-492f-4bee-8b4c-eef4f06eeae2" containerName="ceilometer-central-agent" containerID="cri-o://9f5054fa5f3cb566f03ee591a1ff9f0d5101dc1d2a72a750987d27b7c844b785" gracePeriod=30 Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.196047 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e1922e1-492f-4bee-8b4c-eef4f06eeae2" containerName="sg-core" containerID="cri-o://4e84eada7e930987ecb2f68c26a9194295aa68b506792299af08d2ba022cd9c2" gracePeriod=30 Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.196068 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e1922e1-492f-4bee-8b4c-eef4f06eeae2" containerName="proxy-httpd" containerID="cri-o://971e652bbce0431e8968ac6720e5444c04db44ea13f7178b12fa57b109401870" gracePeriod=30 Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.196095 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2e1922e1-492f-4bee-8b4c-eef4f06eeae2" containerName="ceilometer-notification-agent" containerID="cri-o://1b6b37c96e9bec2f6472cee4d4e32416786dd67b4f3dd297845e0fc86137be2f" gracePeriod=30 Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.597188 4894 generic.go:334] "Generic (PLEG): container finished" podID="2e1922e1-492f-4bee-8b4c-eef4f06eeae2" containerID="971e652bbce0431e8968ac6720e5444c04db44ea13f7178b12fa57b109401870" exitCode=0 Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.597230 4894 generic.go:334] "Generic (PLEG): container finished" podID="2e1922e1-492f-4bee-8b4c-eef4f06eeae2" containerID="4e84eada7e930987ecb2f68c26a9194295aa68b506792299af08d2ba022cd9c2" exitCode=2 Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.597278 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e1922e1-492f-4bee-8b4c-eef4f06eeae2","Type":"ContainerDied","Data":"971e652bbce0431e8968ac6720e5444c04db44ea13f7178b12fa57b109401870"} Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.597310 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e1922e1-492f-4bee-8b4c-eef4f06eeae2","Type":"ContainerDied","Data":"4e84eada7e930987ecb2f68c26a9194295aa68b506792299af08d2ba022cd9c2"} Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.599375 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"3752bfef-7b4f-4f5d-a449-a8b0819f6434","Type":"ContainerStarted","Data":"3b148834876ef8992059219ab1ebf3b53b59184bfa34537b8a3ab2914a4bfe45"} Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.612300 4894 generic.go:334] "Generic (PLEG): container finished" podID="0e41ac7e-15c8-4a9e-ae96-2e68b055ea16" containerID="b2f96b6acf368b01f8d4a3be9093d23033af84a3075e278756e462ebdc719a8b" exitCode=137 Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.612620 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-56f65bfbb-tv5ld" event={"ID":"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16","Type":"ContainerDied","Data":"b2f96b6acf368b01f8d4a3be9093d23033af84a3075e278756e462ebdc719a8b"} Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.629578 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=2.932160263 podStartE2EDuration="14.62955792s" podCreationTimestamp="2025-12-09 15:50:19 +0000 UTC" firstStartedPulling="2025-12-09 15:50:20.63926012 +0000 UTC m=+1114.958470789" lastFinishedPulling="2025-12-09 15:50:32.336657777 +0000 UTC m=+1126.655868446" observedRunningTime="2025-12-09 15:50:33.62126822 +0000 UTC m=+1127.940478889" watchObservedRunningTime="2025-12-09 15:50:33.62955792 +0000 UTC m=+1127.948768579" Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.705406 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.852681 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-config-data\") pod \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.852750 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d65gh\" (UniqueName: \"kubernetes.io/projected/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-kube-api-access-d65gh\") pod \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.852785 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-combined-ca-bundle\") pod \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.852816 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-horizon-secret-key\") pod \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.852844 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-scripts\") pod \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.853023 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-horizon-tls-certs\") pod \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.853115 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-logs\") pod \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\" (UID: \"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16\") " Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.853927 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-logs" (OuterVolumeSpecName: "logs") pod "0e41ac7e-15c8-4a9e-ae96-2e68b055ea16" (UID: "0e41ac7e-15c8-4a9e-ae96-2e68b055ea16"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.864779 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-kube-api-access-d65gh" (OuterVolumeSpecName: "kube-api-access-d65gh") pod "0e41ac7e-15c8-4a9e-ae96-2e68b055ea16" (UID: "0e41ac7e-15c8-4a9e-ae96-2e68b055ea16"). InnerVolumeSpecName "kube-api-access-d65gh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.879871 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-horizon-secret-key" (OuterVolumeSpecName: "horizon-secret-key") pod "0e41ac7e-15c8-4a9e-ae96-2e68b055ea16" (UID: "0e41ac7e-15c8-4a9e-ae96-2e68b055ea16"). InnerVolumeSpecName "horizon-secret-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.880784 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-scripts" (OuterVolumeSpecName: "scripts") pod "0e41ac7e-15c8-4a9e-ae96-2e68b055ea16" (UID: "0e41ac7e-15c8-4a9e-ae96-2e68b055ea16"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.886711 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0e41ac7e-15c8-4a9e-ae96-2e68b055ea16" (UID: "0e41ac7e-15c8-4a9e-ae96-2e68b055ea16"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.903299 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-config-data" (OuterVolumeSpecName: "config-data") pod "0e41ac7e-15c8-4a9e-ae96-2e68b055ea16" (UID: "0e41ac7e-15c8-4a9e-ae96-2e68b055ea16"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.910525 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-horizon-tls-certs" (OuterVolumeSpecName: "horizon-tls-certs") pod "0e41ac7e-15c8-4a9e-ae96-2e68b055ea16" (UID: "0e41ac7e-15c8-4a9e-ae96-2e68b055ea16"). InnerVolumeSpecName "horizon-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.957082 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-logs\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.957118 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.957134 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d65gh\" (UniqueName: \"kubernetes.io/projected/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-kube-api-access-d65gh\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.957147 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.957159 4894 reconciler_common.go:293] "Volume detached for volume \"horizon-secret-key\" (UniqueName: \"kubernetes.io/secret/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-horizon-secret-key\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.957168 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:33 crc kubenswrapper[4894]: I1209 15:50:33.957176 4894 reconciler_common.go:293] "Volume detached for volume \"horizon-tls-certs\" (UniqueName: \"kubernetes.io/secret/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16-horizon-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:34 crc kubenswrapper[4894]: I1209 15:50:34.624991 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizon-56f65bfbb-tv5ld" event={"ID":"0e41ac7e-15c8-4a9e-ae96-2e68b055ea16","Type":"ContainerDied","Data":"8f1e8796e95ab32bb17dd5588ab21f261474d4ba2077138a163f5916d2f31f37"} Dec 09 15:50:34 crc kubenswrapper[4894]: I1209 15:50:34.625090 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizon-56f65bfbb-tv5ld" Dec 09 15:50:34 crc kubenswrapper[4894]: I1209 15:50:34.625378 4894 scope.go:117] "RemoveContainer" containerID="1b3e955a174036764325c570292b5dde1384ef1df60f2868c564915e82f08177" Dec 09 15:50:34 crc kubenswrapper[4894]: I1209 15:50:34.630931 4894 generic.go:334] "Generic (PLEG): container finished" podID="2e1922e1-492f-4bee-8b4c-eef4f06eeae2" containerID="9f5054fa5f3cb566f03ee591a1ff9f0d5101dc1d2a72a750987d27b7c844b785" exitCode=0 Dec 09 15:50:34 crc kubenswrapper[4894]: I1209 15:50:34.631012 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e1922e1-492f-4bee-8b4c-eef4f06eeae2","Type":"ContainerDied","Data":"9f5054fa5f3cb566f03ee591a1ff9f0d5101dc1d2a72a750987d27b7c844b785"} Dec 09 15:50:34 crc kubenswrapper[4894]: I1209 15:50:34.651372 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/horizon-56f65bfbb-tv5ld"] Dec 09 15:50:34 crc kubenswrapper[4894]: I1209 15:50:34.659345 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/horizon-56f65bfbb-tv5ld"] Dec 09 15:50:34 crc kubenswrapper[4894]: I1209 15:50:34.793531 4894 scope.go:117] "RemoveContainer" containerID="b2f96b6acf368b01f8d4a3be9093d23033af84a3075e278756e462ebdc719a8b" Dec 09 15:50:36 crc kubenswrapper[4894]: I1209 15:50:36.118287 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0e41ac7e-15c8-4a9e-ae96-2e68b055ea16" path="/var/lib/kubelet/pods/0e41ac7e-15c8-4a9e-ae96-2e68b055ea16/volumes" Dec 09 15:50:37 crc kubenswrapper[4894]: I1209 15:50:37.669598 4894 generic.go:334] "Generic (PLEG): container finished" podID="2e1922e1-492f-4bee-8b4c-eef4f06eeae2" containerID="1b6b37c96e9bec2f6472cee4d4e32416786dd67b4f3dd297845e0fc86137be2f" exitCode=0 Dec 09 15:50:37 crc kubenswrapper[4894]: I1209 15:50:37.669752 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e1922e1-492f-4bee-8b4c-eef4f06eeae2","Type":"ContainerDied","Data":"1b6b37c96e9bec2f6472cee4d4e32416786dd67b4f3dd297845e0fc86137be2f"} Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.091871 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.251687 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-combined-ca-bundle\") pod \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.251729 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-scripts\") pod \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.251777 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-config-data\") pod \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.251794 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ts7pt\" (UniqueName: \"kubernetes.io/projected/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-kube-api-access-ts7pt\") pod \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.251827 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-run-httpd\") pod \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.251870 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-log-httpd\") pod \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.251916 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-sg-core-conf-yaml\") pod \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\" (UID: \"2e1922e1-492f-4bee-8b4c-eef4f06eeae2\") " Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.252355 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2e1922e1-492f-4bee-8b4c-eef4f06eeae2" (UID: "2e1922e1-492f-4bee-8b4c-eef4f06eeae2"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.252462 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.253308 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2e1922e1-492f-4bee-8b4c-eef4f06eeae2" (UID: "2e1922e1-492f-4bee-8b4c-eef4f06eeae2"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.273414 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-scripts" (OuterVolumeSpecName: "scripts") pod "2e1922e1-492f-4bee-8b4c-eef4f06eeae2" (UID: "2e1922e1-492f-4bee-8b4c-eef4f06eeae2"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.284105 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-kube-api-access-ts7pt" (OuterVolumeSpecName: "kube-api-access-ts7pt") pod "2e1922e1-492f-4bee-8b4c-eef4f06eeae2" (UID: "2e1922e1-492f-4bee-8b4c-eef4f06eeae2"). InnerVolumeSpecName "kube-api-access-ts7pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.297795 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2e1922e1-492f-4bee-8b4c-eef4f06eeae2" (UID: "2e1922e1-492f-4bee-8b4c-eef4f06eeae2"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.331896 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2e1922e1-492f-4bee-8b4c-eef4f06eeae2" (UID: "2e1922e1-492f-4bee-8b4c-eef4f06eeae2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.358625 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.358676 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.358685 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ts7pt\" (UniqueName: \"kubernetes.io/projected/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-kube-api-access-ts7pt\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.358695 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.358729 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.376007 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-config-data" (OuterVolumeSpecName: "config-data") pod "2e1922e1-492f-4bee-8b4c-eef4f06eeae2" (UID: "2e1922e1-492f-4bee-8b4c-eef4f06eeae2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.459959 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2e1922e1-492f-4bee-8b4c-eef4f06eeae2-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.518763 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-r8zsp"] Dec 09 15:50:39 crc kubenswrapper[4894]: E1209 15:50:39.519162 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e41ac7e-15c8-4a9e-ae96-2e68b055ea16" containerName="horizon-log" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.519179 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e41ac7e-15c8-4a9e-ae96-2e68b055ea16" containerName="horizon-log" Dec 09 15:50:39 crc kubenswrapper[4894]: E1209 15:50:39.519194 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e1922e1-492f-4bee-8b4c-eef4f06eeae2" containerName="ceilometer-central-agent" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.519200 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e1922e1-492f-4bee-8b4c-eef4f06eeae2" containerName="ceilometer-central-agent" Dec 09 15:50:39 crc kubenswrapper[4894]: E1209 15:50:39.519214 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e1922e1-492f-4bee-8b4c-eef4f06eeae2" containerName="sg-core" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.519220 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e1922e1-492f-4bee-8b4c-eef4f06eeae2" containerName="sg-core" Dec 09 15:50:39 crc kubenswrapper[4894]: E1209 15:50:39.519231 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e1922e1-492f-4bee-8b4c-eef4f06eeae2" containerName="ceilometer-notification-agent" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.519236 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e1922e1-492f-4bee-8b4c-eef4f06eeae2" containerName="ceilometer-notification-agent" Dec 09 15:50:39 crc kubenswrapper[4894]: E1209 15:50:39.519247 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2e1922e1-492f-4bee-8b4c-eef4f06eeae2" containerName="proxy-httpd" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.519252 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2e1922e1-492f-4bee-8b4c-eef4f06eeae2" containerName="proxy-httpd" Dec 09 15:50:39 crc kubenswrapper[4894]: E1209 15:50:39.519268 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0e41ac7e-15c8-4a9e-ae96-2e68b055ea16" containerName="horizon" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.519274 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0e41ac7e-15c8-4a9e-ae96-2e68b055ea16" containerName="horizon" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.519436 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e1922e1-492f-4bee-8b4c-eef4f06eeae2" containerName="ceilometer-central-agent" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.519448 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e1922e1-492f-4bee-8b4c-eef4f06eeae2" containerName="proxy-httpd" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.519464 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e1922e1-492f-4bee-8b4c-eef4f06eeae2" containerName="sg-core" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.519474 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="2e1922e1-492f-4bee-8b4c-eef4f06eeae2" containerName="ceilometer-notification-agent" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.519481 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e41ac7e-15c8-4a9e-ae96-2e68b055ea16" containerName="horizon-log" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.519496 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0e41ac7e-15c8-4a9e-ae96-2e68b055ea16" containerName="horizon" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.520062 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-r8zsp" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.541133 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-r8zsp"] Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.609419 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-5x62k"] Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.610758 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-5x62k" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.632994 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-5x62k"] Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.648111 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-5951-account-create-update-zl4pj"] Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.649201 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5951-account-create-update-zl4pj" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.653540 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.662944 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xrqgw\" (UniqueName: \"kubernetes.io/projected/ea9963b9-ff7e-4d92-9cf1-7a5108a33b76-kube-api-access-xrqgw\") pod \"nova-api-db-create-r8zsp\" (UID: \"ea9963b9-ff7e-4d92-9cf1-7a5108a33b76\") " pod="openstack/nova-api-db-create-r8zsp" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.662993 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea9963b9-ff7e-4d92-9cf1-7a5108a33b76-operator-scripts\") pod \"nova-api-db-create-r8zsp\" (UID: \"ea9963b9-ff7e-4d92-9cf1-7a5108a33b76\") " pod="openstack/nova-api-db-create-r8zsp" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.672800 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5951-account-create-update-zl4pj"] Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.715877 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2e1922e1-492f-4bee-8b4c-eef4f06eeae2","Type":"ContainerDied","Data":"f1ff300d87ab8014e68a5216b78a266a3cacab1ea59b1417d7209c1731767b41"} Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.716007 4894 scope.go:117] "RemoveContainer" containerID="971e652bbce0431e8968ac6720e5444c04db44ea13f7178b12fa57b109401870" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.716194 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.748146 4894 scope.go:117] "RemoveContainer" containerID="4e84eada7e930987ecb2f68c26a9194295aa68b506792299af08d2ba022cd9c2" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.756610 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.770910 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.772371 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e66b850-9ac9-43ce-977a-a8f61c80a2da-operator-scripts\") pod \"nova-cell0-db-create-5x62k\" (UID: \"8e66b850-9ac9-43ce-977a-a8f61c80a2da\") " pod="openstack/nova-cell0-db-create-5x62k" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.772509 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j92dk\" (UniqueName: \"kubernetes.io/projected/3baa21af-058e-41cc-832b-6e6b138a33d2-kube-api-access-j92dk\") pod \"nova-api-5951-account-create-update-zl4pj\" (UID: \"3baa21af-058e-41cc-832b-6e6b138a33d2\") " pod="openstack/nova-api-5951-account-create-update-zl4pj" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.772575 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3baa21af-058e-41cc-832b-6e6b138a33d2-operator-scripts\") pod \"nova-api-5951-account-create-update-zl4pj\" (UID: \"3baa21af-058e-41cc-832b-6e6b138a33d2\") " pod="openstack/nova-api-5951-account-create-update-zl4pj" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.772698 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fd97k\" (UniqueName: \"kubernetes.io/projected/8e66b850-9ac9-43ce-977a-a8f61c80a2da-kube-api-access-fd97k\") pod \"nova-cell0-db-create-5x62k\" (UID: \"8e66b850-9ac9-43ce-977a-a8f61c80a2da\") " pod="openstack/nova-cell0-db-create-5x62k" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.772762 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xrqgw\" (UniqueName: \"kubernetes.io/projected/ea9963b9-ff7e-4d92-9cf1-7a5108a33b76-kube-api-access-xrqgw\") pod \"nova-api-db-create-r8zsp\" (UID: \"ea9963b9-ff7e-4d92-9cf1-7a5108a33b76\") " pod="openstack/nova-api-db-create-r8zsp" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.772794 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea9963b9-ff7e-4d92-9cf1-7a5108a33b76-operator-scripts\") pod \"nova-api-db-create-r8zsp\" (UID: \"ea9963b9-ff7e-4d92-9cf1-7a5108a33b76\") " pod="openstack/nova-api-db-create-r8zsp" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.773744 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea9963b9-ff7e-4d92-9cf1-7a5108a33b76-operator-scripts\") pod \"nova-api-db-create-r8zsp\" (UID: \"ea9963b9-ff7e-4d92-9cf1-7a5108a33b76\") " pod="openstack/nova-api-db-create-r8zsp" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.790205 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.795769 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.801782 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.801887 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.807535 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xrqgw\" (UniqueName: \"kubernetes.io/projected/ea9963b9-ff7e-4d92-9cf1-7a5108a33b76-kube-api-access-xrqgw\") pod \"nova-api-db-create-r8zsp\" (UID: \"ea9963b9-ff7e-4d92-9cf1-7a5108a33b76\") " pod="openstack/nova-api-db-create-r8zsp" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.811530 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.815236 4894 scope.go:117] "RemoveContainer" containerID="1b6b37c96e9bec2f6472cee4d4e32416786dd67b4f3dd297845e0fc86137be2f" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.831315 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-vt68w"] Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.832788 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-vt68w" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.836517 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-r8zsp" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.840392 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-vt68w"] Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.850714 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-9b21-account-create-update-g5mj7"] Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.850785 4894 scope.go:117] "RemoveContainer" containerID="9f5054fa5f3cb566f03ee591a1ff9f0d5101dc1d2a72a750987d27b7c844b785" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.853694 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9b21-account-create-update-g5mj7" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.855237 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.877745 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9b21-account-create-update-g5mj7"] Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.877845 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j92dk\" (UniqueName: \"kubernetes.io/projected/3baa21af-058e-41cc-832b-6e6b138a33d2-kube-api-access-j92dk\") pod \"nova-api-5951-account-create-update-zl4pj\" (UID: \"3baa21af-058e-41cc-832b-6e6b138a33d2\") " pod="openstack/nova-api-5951-account-create-update-zl4pj" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.877899 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3baa21af-058e-41cc-832b-6e6b138a33d2-operator-scripts\") pod \"nova-api-5951-account-create-update-zl4pj\" (UID: \"3baa21af-058e-41cc-832b-6e6b138a33d2\") " pod="openstack/nova-api-5951-account-create-update-zl4pj" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.878006 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fd97k\" (UniqueName: \"kubernetes.io/projected/8e66b850-9ac9-43ce-977a-a8f61c80a2da-kube-api-access-fd97k\") pod \"nova-cell0-db-create-5x62k\" (UID: \"8e66b850-9ac9-43ce-977a-a8f61c80a2da\") " pod="openstack/nova-cell0-db-create-5x62k" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.878147 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e66b850-9ac9-43ce-977a-a8f61c80a2da-operator-scripts\") pod \"nova-cell0-db-create-5x62k\" (UID: \"8e66b850-9ac9-43ce-977a-a8f61c80a2da\") " pod="openstack/nova-cell0-db-create-5x62k" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.879033 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3baa21af-058e-41cc-832b-6e6b138a33d2-operator-scripts\") pod \"nova-api-5951-account-create-update-zl4pj\" (UID: \"3baa21af-058e-41cc-832b-6e6b138a33d2\") " pod="openstack/nova-api-5951-account-create-update-zl4pj" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.879133 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e66b850-9ac9-43ce-977a-a8f61c80a2da-operator-scripts\") pod \"nova-cell0-db-create-5x62k\" (UID: \"8e66b850-9ac9-43ce-977a-a8f61c80a2da\") " pod="openstack/nova-cell0-db-create-5x62k" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.902158 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j92dk\" (UniqueName: \"kubernetes.io/projected/3baa21af-058e-41cc-832b-6e6b138a33d2-kube-api-access-j92dk\") pod \"nova-api-5951-account-create-update-zl4pj\" (UID: \"3baa21af-058e-41cc-832b-6e6b138a33d2\") " pod="openstack/nova-api-5951-account-create-update-zl4pj" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.905220 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fd97k\" (UniqueName: \"kubernetes.io/projected/8e66b850-9ac9-43ce-977a-a8f61c80a2da-kube-api-access-fd97k\") pod \"nova-cell0-db-create-5x62k\" (UID: \"8e66b850-9ac9-43ce-977a-a8f61c80a2da\") " pod="openstack/nova-cell0-db-create-5x62k" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.907444 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:50:39 crc kubenswrapper[4894]: E1209 15:50:39.908207 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[combined-ca-bundle config-data kube-api-access-tpsf5 log-httpd run-httpd scripts sg-core-conf-yaml], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/ceilometer-0" podUID="de2710f5-d1af-4987-9f58-80846fb85e77" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.952459 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-5x62k" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.969216 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5951-account-create-update-zl4pj" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.980603 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " pod="openstack/ceilometer-0" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.980725 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mzdkj\" (UniqueName: \"kubernetes.io/projected/22fdb6d9-4858-44bf-a856-04b66398260f-kube-api-access-mzdkj\") pod \"nova-cell0-9b21-account-create-update-g5mj7\" (UID: \"22fdb6d9-4858-44bf-a856-04b66398260f\") " pod="openstack/nova-cell0-9b21-account-create-update-g5mj7" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.980752 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-config-data\") pod \"ceilometer-0\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " pod="openstack/ceilometer-0" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.980772 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de2710f5-d1af-4987-9f58-80846fb85e77-run-httpd\") pod \"ceilometer-0\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " pod="openstack/ceilometer-0" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.980851 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22fdb6d9-4858-44bf-a856-04b66398260f-operator-scripts\") pod \"nova-cell0-9b21-account-create-update-g5mj7\" (UID: \"22fdb6d9-4858-44bf-a856-04b66398260f\") " pod="openstack/nova-cell0-9b21-account-create-update-g5mj7" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.980912 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-scripts\") pod \"ceilometer-0\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " pod="openstack/ceilometer-0" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.981034 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " pod="openstack/ceilometer-0" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.981058 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de2710f5-d1af-4987-9f58-80846fb85e77-log-httpd\") pod \"ceilometer-0\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " pod="openstack/ceilometer-0" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.981078 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qqvvt\" (UniqueName: \"kubernetes.io/projected/88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c-kube-api-access-qqvvt\") pod \"nova-cell1-db-create-vt68w\" (UID: \"88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c\") " pod="openstack/nova-cell1-db-create-vt68w" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.981097 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c-operator-scripts\") pod \"nova-cell1-db-create-vt68w\" (UID: \"88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c\") " pod="openstack/nova-cell1-db-create-vt68w" Dec 09 15:50:39 crc kubenswrapper[4894]: I1209 15:50:39.981153 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tpsf5\" (UniqueName: \"kubernetes.io/projected/de2710f5-d1af-4987-9f58-80846fb85e77-kube-api-access-tpsf5\") pod \"ceilometer-0\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " pod="openstack/ceilometer-0" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.040738 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-9b51-account-create-update-xvkxq"] Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.045327 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9b51-account-create-update-xvkxq" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.047426 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.050857 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-9b51-account-create-update-xvkxq"] Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.083592 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mzdkj\" (UniqueName: \"kubernetes.io/projected/22fdb6d9-4858-44bf-a856-04b66398260f-kube-api-access-mzdkj\") pod \"nova-cell0-9b21-account-create-update-g5mj7\" (UID: \"22fdb6d9-4858-44bf-a856-04b66398260f\") " pod="openstack/nova-cell0-9b21-account-create-update-g5mj7" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.083662 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-config-data\") pod \"ceilometer-0\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " pod="openstack/ceilometer-0" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.083694 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de2710f5-d1af-4987-9f58-80846fb85e77-run-httpd\") pod \"ceilometer-0\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " pod="openstack/ceilometer-0" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.083729 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22fdb6d9-4858-44bf-a856-04b66398260f-operator-scripts\") pod \"nova-cell0-9b21-account-create-update-g5mj7\" (UID: \"22fdb6d9-4858-44bf-a856-04b66398260f\") " pod="openstack/nova-cell0-9b21-account-create-update-g5mj7" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.083761 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-scripts\") pod \"ceilometer-0\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " pod="openstack/ceilometer-0" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.083844 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " pod="openstack/ceilometer-0" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.083872 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qqvvt\" (UniqueName: \"kubernetes.io/projected/88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c-kube-api-access-qqvvt\") pod \"nova-cell1-db-create-vt68w\" (UID: \"88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c\") " pod="openstack/nova-cell1-db-create-vt68w" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.083891 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de2710f5-d1af-4987-9f58-80846fb85e77-log-httpd\") pod \"ceilometer-0\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " pod="openstack/ceilometer-0" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.083914 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c-operator-scripts\") pod \"nova-cell1-db-create-vt68w\" (UID: \"88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c\") " pod="openstack/nova-cell1-db-create-vt68w" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.083962 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tpsf5\" (UniqueName: \"kubernetes.io/projected/de2710f5-d1af-4987-9f58-80846fb85e77-kube-api-access-tpsf5\") pod \"ceilometer-0\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " pod="openstack/ceilometer-0" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.084019 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " pod="openstack/ceilometer-0" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.085446 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de2710f5-d1af-4987-9f58-80846fb85e77-run-httpd\") pod \"ceilometer-0\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " pod="openstack/ceilometer-0" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.085700 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de2710f5-d1af-4987-9f58-80846fb85e77-log-httpd\") pod \"ceilometer-0\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " pod="openstack/ceilometer-0" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.086331 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c-operator-scripts\") pod \"nova-cell1-db-create-vt68w\" (UID: \"88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c\") " pod="openstack/nova-cell1-db-create-vt68w" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.086957 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22fdb6d9-4858-44bf-a856-04b66398260f-operator-scripts\") pod \"nova-cell0-9b21-account-create-update-g5mj7\" (UID: \"22fdb6d9-4858-44bf-a856-04b66398260f\") " pod="openstack/nova-cell0-9b21-account-create-update-g5mj7" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.104093 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " pod="openstack/ceilometer-0" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.105322 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-scripts\") pod \"ceilometer-0\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " pod="openstack/ceilometer-0" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.120960 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " pod="openstack/ceilometer-0" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.131502 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qqvvt\" (UniqueName: \"kubernetes.io/projected/88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c-kube-api-access-qqvvt\") pod \"nova-cell1-db-create-vt68w\" (UID: \"88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c\") " pod="openstack/nova-cell1-db-create-vt68w" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.131723 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mzdkj\" (UniqueName: \"kubernetes.io/projected/22fdb6d9-4858-44bf-a856-04b66398260f-kube-api-access-mzdkj\") pod \"nova-cell0-9b21-account-create-update-g5mj7\" (UID: \"22fdb6d9-4858-44bf-a856-04b66398260f\") " pod="openstack/nova-cell0-9b21-account-create-update-g5mj7" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.141536 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tpsf5\" (UniqueName: \"kubernetes.io/projected/de2710f5-d1af-4987-9f58-80846fb85e77-kube-api-access-tpsf5\") pod \"ceilometer-0\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " pod="openstack/ceilometer-0" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.155167 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2e1922e1-492f-4bee-8b4c-eef4f06eeae2" path="/var/lib/kubelet/pods/2e1922e1-492f-4bee-8b4c-eef4f06eeae2/volumes" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.156979 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-config-data\") pod \"ceilometer-0\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " pod="openstack/ceilometer-0" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.181087 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-vt68w" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.187068 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/423a2183-9395-4c3f-ab1c-1816c1d31fb9-operator-scripts\") pod \"nova-cell1-9b51-account-create-update-xvkxq\" (UID: \"423a2183-9395-4c3f-ab1c-1816c1d31fb9\") " pod="openstack/nova-cell1-9b51-account-create-update-xvkxq" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.187646 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j79fm\" (UniqueName: \"kubernetes.io/projected/423a2183-9395-4c3f-ab1c-1816c1d31fb9-kube-api-access-j79fm\") pod \"nova-cell1-9b51-account-create-update-xvkxq\" (UID: \"423a2183-9395-4c3f-ab1c-1816c1d31fb9\") " pod="openstack/nova-cell1-9b51-account-create-update-xvkxq" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.200920 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9b21-account-create-update-g5mj7" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.290984 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/423a2183-9395-4c3f-ab1c-1816c1d31fb9-operator-scripts\") pod \"nova-cell1-9b51-account-create-update-xvkxq\" (UID: \"423a2183-9395-4c3f-ab1c-1816c1d31fb9\") " pod="openstack/nova-cell1-9b51-account-create-update-xvkxq" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.291331 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j79fm\" (UniqueName: \"kubernetes.io/projected/423a2183-9395-4c3f-ab1c-1816c1d31fb9-kube-api-access-j79fm\") pod \"nova-cell1-9b51-account-create-update-xvkxq\" (UID: \"423a2183-9395-4c3f-ab1c-1816c1d31fb9\") " pod="openstack/nova-cell1-9b51-account-create-update-xvkxq" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.293306 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/423a2183-9395-4c3f-ab1c-1816c1d31fb9-operator-scripts\") pod \"nova-cell1-9b51-account-create-update-xvkxq\" (UID: \"423a2183-9395-4c3f-ab1c-1816c1d31fb9\") " pod="openstack/nova-cell1-9b51-account-create-update-xvkxq" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.316580 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j79fm\" (UniqueName: \"kubernetes.io/projected/423a2183-9395-4c3f-ab1c-1816c1d31fb9-kube-api-access-j79fm\") pod \"nova-cell1-9b51-account-create-update-xvkxq\" (UID: \"423a2183-9395-4c3f-ab1c-1816c1d31fb9\") " pod="openstack/nova-cell1-9b51-account-create-update-xvkxq" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.406692 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-5951-account-create-update-zl4pj"] Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.419919 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-r8zsp"] Dec 09 15:50:40 crc kubenswrapper[4894]: W1209 15:50:40.426660 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podea9963b9_ff7e_4d92_9cf1_7a5108a33b76.slice/crio-5f00c551a470b3a0d6b8178542409ce83c2fcf39c8b411325b15d96262c46fff WatchSource:0}: Error finding container 5f00c551a470b3a0d6b8178542409ce83c2fcf39c8b411325b15d96262c46fff: Status 404 returned error can't find the container with id 5f00c551a470b3a0d6b8178542409ce83c2fcf39c8b411325b15d96262c46fff Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.541088 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-5x62k"] Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.554180 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9b51-account-create-update-xvkxq" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.730966 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5951-account-create-update-zl4pj" event={"ID":"3baa21af-058e-41cc-832b-6e6b138a33d2","Type":"ContainerStarted","Data":"21e1648dc1dccfed5f68bef50d540de6507f298f0a3b91b0fe33cf916a487e4f"} Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.731018 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5951-account-create-update-zl4pj" event={"ID":"3baa21af-058e-41cc-832b-6e6b138a33d2","Type":"ContainerStarted","Data":"3d4cfdda7318b24caf98c0e9936c15803ae1af2ebf58e9e89a9dc23db0f3d1c1"} Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.734961 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-r8zsp" event={"ID":"ea9963b9-ff7e-4d92-9cf1-7a5108a33b76","Type":"ContainerStarted","Data":"6b95ed24650ca49fa2e0af023c8af9711bcfa64f731d350e210e70cc20b97e37"} Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.734993 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-r8zsp" event={"ID":"ea9963b9-ff7e-4d92-9cf1-7a5108a33b76","Type":"ContainerStarted","Data":"5f00c551a470b3a0d6b8178542409ce83c2fcf39c8b411325b15d96262c46fff"} Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.762854 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.762965 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-5x62k" event={"ID":"8e66b850-9ac9-43ce-977a-a8f61c80a2da","Type":"ContainerStarted","Data":"19cca9c07ba48a5f43b63d2d9f6068a2048d5c05390daa973c00896a6e47eed3"} Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.763011 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-5x62k" event={"ID":"8e66b850-9ac9-43ce-977a-a8f61c80a2da","Type":"ContainerStarted","Data":"fe07168589d5548992764db66a4d1ed7fb2cd8ab123239643ecffeab01984502"} Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.765213 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-5951-account-create-update-zl4pj" podStartSLOduration=1.765194765 podStartE2EDuration="1.765194765s" podCreationTimestamp="2025-12-09 15:50:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:50:40.753402117 +0000 UTC m=+1135.072612786" watchObservedRunningTime="2025-12-09 15:50:40.765194765 +0000 UTC m=+1135.084405434" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.775649 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.780801 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-9b21-account-create-update-g5mj7"] Dec 09 15:50:40 crc kubenswrapper[4894]: W1209 15:50:40.782962 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod22fdb6d9_4858_44bf_a856_04b66398260f.slice/crio-a92f04bf213953a71361f2ccf669889dd26cbbc7a9980cb6e831146a23937e3f WatchSource:0}: Error finding container a92f04bf213953a71361f2ccf669889dd26cbbc7a9980cb6e831146a23937e3f: Status 404 returned error can't find the container with id a92f04bf213953a71361f2ccf669889dd26cbbc7a9980cb6e831146a23937e3f Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.800603 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-vt68w"] Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.806281 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-r8zsp" podStartSLOduration=1.806260075 podStartE2EDuration="1.806260075s" podCreationTimestamp="2025-12-09 15:50:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:50:40.777699012 +0000 UTC m=+1135.096909691" watchObservedRunningTime="2025-12-09 15:50:40.806260075 +0000 UTC m=+1135.125470744" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.816176 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-db-create-5x62k" podStartSLOduration=1.816157945 podStartE2EDuration="1.816157945s" podCreationTimestamp="2025-12-09 15:50:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:50:40.797011071 +0000 UTC m=+1135.116221740" watchObservedRunningTime="2025-12-09 15:50:40.816157945 +0000 UTC m=+1135.135368614" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.902476 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-config-data\") pod \"de2710f5-d1af-4987-9f58-80846fb85e77\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.902571 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de2710f5-d1af-4987-9f58-80846fb85e77-log-httpd\") pod \"de2710f5-d1af-4987-9f58-80846fb85e77\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.902589 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-combined-ca-bundle\") pod \"de2710f5-d1af-4987-9f58-80846fb85e77\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.902713 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de2710f5-d1af-4987-9f58-80846fb85e77-run-httpd\") pod \"de2710f5-d1af-4987-9f58-80846fb85e77\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.902748 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-scripts\") pod \"de2710f5-d1af-4987-9f58-80846fb85e77\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.902805 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tpsf5\" (UniqueName: \"kubernetes.io/projected/de2710f5-d1af-4987-9f58-80846fb85e77-kube-api-access-tpsf5\") pod \"de2710f5-d1af-4987-9f58-80846fb85e77\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.902853 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-sg-core-conf-yaml\") pod \"de2710f5-d1af-4987-9f58-80846fb85e77\" (UID: \"de2710f5-d1af-4987-9f58-80846fb85e77\") " Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.903500 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de2710f5-d1af-4987-9f58-80846fb85e77-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "de2710f5-d1af-4987-9f58-80846fb85e77" (UID: "de2710f5-d1af-4987-9f58-80846fb85e77"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.903886 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/de2710f5-d1af-4987-9f58-80846fb85e77-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "de2710f5-d1af-4987-9f58-80846fb85e77" (UID: "de2710f5-d1af-4987-9f58-80846fb85e77"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.908806 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-scripts" (OuterVolumeSpecName: "scripts") pod "de2710f5-d1af-4987-9f58-80846fb85e77" (UID: "de2710f5-d1af-4987-9f58-80846fb85e77"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.908860 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de2710f5-d1af-4987-9f58-80846fb85e77-kube-api-access-tpsf5" (OuterVolumeSpecName: "kube-api-access-tpsf5") pod "de2710f5-d1af-4987-9f58-80846fb85e77" (UID: "de2710f5-d1af-4987-9f58-80846fb85e77"). InnerVolumeSpecName "kube-api-access-tpsf5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.909796 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "de2710f5-d1af-4987-9f58-80846fb85e77" (UID: "de2710f5-d1af-4987-9f58-80846fb85e77"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.910773 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "de2710f5-d1af-4987-9f58-80846fb85e77" (UID: "de2710f5-d1af-4987-9f58-80846fb85e77"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:40 crc kubenswrapper[4894]: I1209 15:50:40.910898 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-config-data" (OuterVolumeSpecName: "config-data") pod "de2710f5-d1af-4987-9f58-80846fb85e77" (UID: "de2710f5-d1af-4987-9f58-80846fb85e77"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:50:41 crc kubenswrapper[4894]: I1209 15:50:41.018207 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:41 crc kubenswrapper[4894]: I1209 15:50:41.018542 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tpsf5\" (UniqueName: \"kubernetes.io/projected/de2710f5-d1af-4987-9f58-80846fb85e77-kube-api-access-tpsf5\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:41 crc kubenswrapper[4894]: I1209 15:50:41.018556 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:41 crc kubenswrapper[4894]: I1209 15:50:41.018566 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:41 crc kubenswrapper[4894]: I1209 15:50:41.018577 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de2710f5-d1af-4987-9f58-80846fb85e77-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:41 crc kubenswrapper[4894]: I1209 15:50:41.018587 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/de2710f5-d1af-4987-9f58-80846fb85e77-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:41 crc kubenswrapper[4894]: I1209 15:50:41.018599 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/de2710f5-d1af-4987-9f58-80846fb85e77-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:41 crc kubenswrapper[4894]: I1209 15:50:41.073435 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-9b51-account-create-update-xvkxq"] Dec 09 15:50:41 crc kubenswrapper[4894]: W1209 15:50:41.077477 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod423a2183_9395_4c3f_ab1c_1816c1d31fb9.slice/crio-dcf8f20731f814fbed4ffc86bb88fd6ca61da2e9240546c0720fbd6637e72817 WatchSource:0}: Error finding container dcf8f20731f814fbed4ffc86bb88fd6ca61da2e9240546c0720fbd6637e72817: Status 404 returned error can't find the container with id dcf8f20731f814fbed4ffc86bb88fd6ca61da2e9240546c0720fbd6637e72817 Dec 09 15:50:41 crc kubenswrapper[4894]: I1209 15:50:41.771254 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-vt68w" event={"ID":"88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c","Type":"ContainerStarted","Data":"8eadbc597a691bb8c6f1de7c942e49a967878dbf3416d1b7fb58f127709103c5"} Dec 09 15:50:41 crc kubenswrapper[4894]: I1209 15:50:41.772577 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9b21-account-create-update-g5mj7" event={"ID":"22fdb6d9-4858-44bf-a856-04b66398260f","Type":"ContainerStarted","Data":"a92f04bf213953a71361f2ccf669889dd26cbbc7a9980cb6e831146a23937e3f"} Dec 09 15:50:41 crc kubenswrapper[4894]: I1209 15:50:41.773682 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9b51-account-create-update-xvkxq" event={"ID":"423a2183-9395-4c3f-ab1c-1816c1d31fb9","Type":"ContainerStarted","Data":"dcf8f20731f814fbed4ffc86bb88fd6ca61da2e9240546c0720fbd6637e72817"} Dec 09 15:50:41 crc kubenswrapper[4894]: I1209 15:50:41.775518 4894 generic.go:334] "Generic (PLEG): container finished" podID="ea9963b9-ff7e-4d92-9cf1-7a5108a33b76" containerID="6b95ed24650ca49fa2e0af023c8af9711bcfa64f731d350e210e70cc20b97e37" exitCode=0 Dec 09 15:50:41 crc kubenswrapper[4894]: I1209 15:50:41.775556 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-r8zsp" event={"ID":"ea9963b9-ff7e-4d92-9cf1-7a5108a33b76","Type":"ContainerDied","Data":"6b95ed24650ca49fa2e0af023c8af9711bcfa64f731d350e210e70cc20b97e37"} Dec 09 15:50:41 crc kubenswrapper[4894]: I1209 15:50:41.775621 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:50:41 crc kubenswrapper[4894]: I1209 15:50:41.845932 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:50:41 crc kubenswrapper[4894]: I1209 15:50:41.874041 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:50:41 crc kubenswrapper[4894]: I1209 15:50:41.887809 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:50:41 crc kubenswrapper[4894]: I1209 15:50:41.890013 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:50:41 crc kubenswrapper[4894]: I1209 15:50:41.892719 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 15:50:41 crc kubenswrapper[4894]: I1209 15:50:41.892977 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 15:50:41 crc kubenswrapper[4894]: I1209 15:50:41.902604 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.038922 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbwsd\" (UniqueName: \"kubernetes.io/projected/4f44ed30-5746-4bc4-80b4-53b5622c13ae-kube-api-access-rbwsd\") pod \"ceilometer-0\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " pod="openstack/ceilometer-0" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.039013 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-scripts\") pod \"ceilometer-0\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " pod="openstack/ceilometer-0" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.039075 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " pod="openstack/ceilometer-0" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.039145 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-config-data\") pod \"ceilometer-0\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " pod="openstack/ceilometer-0" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.039177 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " pod="openstack/ceilometer-0" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.039274 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f44ed30-5746-4bc4-80b4-53b5622c13ae-run-httpd\") pod \"ceilometer-0\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " pod="openstack/ceilometer-0" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.039306 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f44ed30-5746-4bc4-80b4-53b5622c13ae-log-httpd\") pod \"ceilometer-0\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " pod="openstack/ceilometer-0" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.116739 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de2710f5-d1af-4987-9f58-80846fb85e77" path="/var/lib/kubelet/pods/de2710f5-d1af-4987-9f58-80846fb85e77/volumes" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.141315 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-scripts\") pod \"ceilometer-0\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " pod="openstack/ceilometer-0" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.141376 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " pod="openstack/ceilometer-0" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.141418 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-config-data\") pod \"ceilometer-0\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " pod="openstack/ceilometer-0" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.141439 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " pod="openstack/ceilometer-0" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.141487 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f44ed30-5746-4bc4-80b4-53b5622c13ae-run-httpd\") pod \"ceilometer-0\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " pod="openstack/ceilometer-0" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.141507 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f44ed30-5746-4bc4-80b4-53b5622c13ae-log-httpd\") pod \"ceilometer-0\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " pod="openstack/ceilometer-0" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.141543 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbwsd\" (UniqueName: \"kubernetes.io/projected/4f44ed30-5746-4bc4-80b4-53b5622c13ae-kube-api-access-rbwsd\") pod \"ceilometer-0\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " pod="openstack/ceilometer-0" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.142463 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f44ed30-5746-4bc4-80b4-53b5622c13ae-run-httpd\") pod \"ceilometer-0\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " pod="openstack/ceilometer-0" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.142535 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f44ed30-5746-4bc4-80b4-53b5622c13ae-log-httpd\") pod \"ceilometer-0\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " pod="openstack/ceilometer-0" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.147198 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " pod="openstack/ceilometer-0" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.147240 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-scripts\") pod \"ceilometer-0\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " pod="openstack/ceilometer-0" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.147584 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-config-data\") pod \"ceilometer-0\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " pod="openstack/ceilometer-0" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.148070 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " pod="openstack/ceilometer-0" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.164053 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbwsd\" (UniqueName: \"kubernetes.io/projected/4f44ed30-5746-4bc4-80b4-53b5622c13ae-kube-api-access-rbwsd\") pod \"ceilometer-0\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " pod="openstack/ceilometer-0" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.217669 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.257337 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.257557 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.257871 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.258906 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"06d076fbf334b50ab934fbbefedcfc6bb4d75c33f4047f53bb661bb7b87d68ea"} pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.259177 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" containerID="cri-o://06d076fbf334b50ab934fbbefedcfc6bb4d75c33f4047f53bb661bb7b87d68ea" gracePeriod=600 Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.670592 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:50:42 crc kubenswrapper[4894]: W1209 15:50:42.674317 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4f44ed30_5746_4bc4_80b4_53b5622c13ae.slice/crio-156a48ba07176f332f47253915b8f7d179afeec2b7a8db2fcbb0bce90e27bd83 WatchSource:0}: Error finding container 156a48ba07176f332f47253915b8f7d179afeec2b7a8db2fcbb0bce90e27bd83: Status 404 returned error can't find the container with id 156a48ba07176f332f47253915b8f7d179afeec2b7a8db2fcbb0bce90e27bd83 Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.784813 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9b21-account-create-update-g5mj7" event={"ID":"22fdb6d9-4858-44bf-a856-04b66398260f","Type":"ContainerStarted","Data":"ebb2d23f3bfa50765824c94ae14d468a88a6a0b774c71abcf01a4eb038186069"} Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.785931 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f44ed30-5746-4bc4-80b4-53b5622c13ae","Type":"ContainerStarted","Data":"156a48ba07176f332f47253915b8f7d179afeec2b7a8db2fcbb0bce90e27bd83"} Dec 09 15:50:42 crc kubenswrapper[4894]: I1209 15:50:42.791013 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-vt68w" event={"ID":"88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c","Type":"ContainerStarted","Data":"c13c0b98297df1f9d64e1feab246d03958c398c3a45c527d5b51479615e8a04a"} Dec 09 15:50:43 crc kubenswrapper[4894]: I1209 15:50:43.089672 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-r8zsp" Dec 09 15:50:43 crc kubenswrapper[4894]: I1209 15:50:43.262102 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea9963b9-ff7e-4d92-9cf1-7a5108a33b76-operator-scripts\") pod \"ea9963b9-ff7e-4d92-9cf1-7a5108a33b76\" (UID: \"ea9963b9-ff7e-4d92-9cf1-7a5108a33b76\") " Dec 09 15:50:43 crc kubenswrapper[4894]: I1209 15:50:43.262240 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xrqgw\" (UniqueName: \"kubernetes.io/projected/ea9963b9-ff7e-4d92-9cf1-7a5108a33b76-kube-api-access-xrqgw\") pod \"ea9963b9-ff7e-4d92-9cf1-7a5108a33b76\" (UID: \"ea9963b9-ff7e-4d92-9cf1-7a5108a33b76\") " Dec 09 15:50:43 crc kubenswrapper[4894]: I1209 15:50:43.263069 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ea9963b9-ff7e-4d92-9cf1-7a5108a33b76-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ea9963b9-ff7e-4d92-9cf1-7a5108a33b76" (UID: "ea9963b9-ff7e-4d92-9cf1-7a5108a33b76"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:50:43 crc kubenswrapper[4894]: I1209 15:50:43.270938 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea9963b9-ff7e-4d92-9cf1-7a5108a33b76-kube-api-access-xrqgw" (OuterVolumeSpecName: "kube-api-access-xrqgw") pod "ea9963b9-ff7e-4d92-9cf1-7a5108a33b76" (UID: "ea9963b9-ff7e-4d92-9cf1-7a5108a33b76"). InnerVolumeSpecName "kube-api-access-xrqgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:50:43 crc kubenswrapper[4894]: I1209 15:50:43.364389 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ea9963b9-ff7e-4d92-9cf1-7a5108a33b76-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:43 crc kubenswrapper[4894]: I1209 15:50:43.364432 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xrqgw\" (UniqueName: \"kubernetes.io/projected/ea9963b9-ff7e-4d92-9cf1-7a5108a33b76-kube-api-access-xrqgw\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:43 crc kubenswrapper[4894]: I1209 15:50:43.807016 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9b51-account-create-update-xvkxq" event={"ID":"423a2183-9395-4c3f-ab1c-1816c1d31fb9","Type":"ContainerStarted","Data":"a622ced4830d636677017d89b88e3a2170266e342db1cbe48a5537d668c423cd"} Dec 09 15:50:43 crc kubenswrapper[4894]: I1209 15:50:43.809914 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-r8zsp" event={"ID":"ea9963b9-ff7e-4d92-9cf1-7a5108a33b76","Type":"ContainerDied","Data":"5f00c551a470b3a0d6b8178542409ce83c2fcf39c8b411325b15d96262c46fff"} Dec 09 15:50:43 crc kubenswrapper[4894]: I1209 15:50:43.809976 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5f00c551a470b3a0d6b8178542409ce83c2fcf39c8b411325b15d96262c46fff" Dec 09 15:50:43 crc kubenswrapper[4894]: I1209 15:50:43.810082 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-r8zsp" Dec 09 15:50:49 crc kubenswrapper[4894]: I1209 15:50:49.541996 4894 generic.go:334] "Generic (PLEG): container finished" podID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerID="06d076fbf334b50ab934fbbefedcfc6bb4d75c33f4047f53bb661bb7b87d68ea" exitCode=0 Dec 09 15:50:49 crc kubenswrapper[4894]: I1209 15:50:49.542045 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerDied","Data":"06d076fbf334b50ab934fbbefedcfc6bb4d75c33f4047f53bb661bb7b87d68ea"} Dec 09 15:50:49 crc kubenswrapper[4894]: I1209 15:50:49.542512 4894 scope.go:117] "RemoveContainer" containerID="3c92ffd29e4b95c91cb30f4577bf2eab4718185324fe428051f5af6db9d8c9a8" Dec 09 15:50:50 crc kubenswrapper[4894]: I1209 15:50:50.554342 4894 generic.go:334] "Generic (PLEG): container finished" podID="88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c" containerID="c13c0b98297df1f9d64e1feab246d03958c398c3a45c527d5b51479615e8a04a" exitCode=0 Dec 09 15:50:50 crc kubenswrapper[4894]: I1209 15:50:50.554447 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-vt68w" event={"ID":"88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c","Type":"ContainerDied","Data":"c13c0b98297df1f9d64e1feab246d03958c398c3a45c527d5b51479615e8a04a"} Dec 09 15:50:50 crc kubenswrapper[4894]: I1209 15:50:50.558255 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerStarted","Data":"cff4b305c5655888c5cc65c71b639fe907b50486a6daafac98221b58e06cb929"} Dec 09 15:50:50 crc kubenswrapper[4894]: I1209 15:50:50.560012 4894 generic.go:334] "Generic (PLEG): container finished" podID="3baa21af-058e-41cc-832b-6e6b138a33d2" containerID="21e1648dc1dccfed5f68bef50d540de6507f298f0a3b91b0fe33cf916a487e4f" exitCode=0 Dec 09 15:50:50 crc kubenswrapper[4894]: I1209 15:50:50.560080 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5951-account-create-update-zl4pj" event={"ID":"3baa21af-058e-41cc-832b-6e6b138a33d2","Type":"ContainerDied","Data":"21e1648dc1dccfed5f68bef50d540de6507f298f0a3b91b0fe33cf916a487e4f"} Dec 09 15:50:50 crc kubenswrapper[4894]: I1209 15:50:50.586629 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-9b51-account-create-update-xvkxq" podStartSLOduration=11.586610178 podStartE2EDuration="11.586610178s" podCreationTimestamp="2025-12-09 15:50:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:50:50.581980601 +0000 UTC m=+1144.901191270" watchObservedRunningTime="2025-12-09 15:50:50.586610178 +0000 UTC m=+1144.905820837" Dec 09 15:50:50 crc kubenswrapper[4894]: I1209 15:50:50.629816 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-9b21-account-create-update-g5mj7" podStartSLOduration=11.629796721 podStartE2EDuration="11.629796721s" podCreationTimestamp="2025-12-09 15:50:39 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:50:50.624329242 +0000 UTC m=+1144.943539911" watchObservedRunningTime="2025-12-09 15:50:50.629796721 +0000 UTC m=+1144.949007390" Dec 09 15:50:51 crc kubenswrapper[4894]: I1209 15:50:51.584749 4894 generic.go:334] "Generic (PLEG): container finished" podID="8e66b850-9ac9-43ce-977a-a8f61c80a2da" containerID="19cca9c07ba48a5f43b63d2d9f6068a2048d5c05390daa973c00896a6e47eed3" exitCode=0 Dec 09 15:50:51 crc kubenswrapper[4894]: I1209 15:50:51.584894 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-5x62k" event={"ID":"8e66b850-9ac9-43ce-977a-a8f61c80a2da","Type":"ContainerDied","Data":"19cca9c07ba48a5f43b63d2d9f6068a2048d5c05390daa973c00896a6e47eed3"} Dec 09 15:50:51 crc kubenswrapper[4894]: I1209 15:50:51.962303 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5951-account-create-update-zl4pj" Dec 09 15:50:51 crc kubenswrapper[4894]: I1209 15:50:51.967106 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-vt68w" Dec 09 15:50:52 crc kubenswrapper[4894]: I1209 15:50:52.122318 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qqvvt\" (UniqueName: \"kubernetes.io/projected/88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c-kube-api-access-qqvvt\") pod \"88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c\" (UID: \"88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c\") " Dec 09 15:50:52 crc kubenswrapper[4894]: I1209 15:50:52.122437 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j92dk\" (UniqueName: \"kubernetes.io/projected/3baa21af-058e-41cc-832b-6e6b138a33d2-kube-api-access-j92dk\") pod \"3baa21af-058e-41cc-832b-6e6b138a33d2\" (UID: \"3baa21af-058e-41cc-832b-6e6b138a33d2\") " Dec 09 15:50:52 crc kubenswrapper[4894]: I1209 15:50:52.122531 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3baa21af-058e-41cc-832b-6e6b138a33d2-operator-scripts\") pod \"3baa21af-058e-41cc-832b-6e6b138a33d2\" (UID: \"3baa21af-058e-41cc-832b-6e6b138a33d2\") " Dec 09 15:50:52 crc kubenswrapper[4894]: I1209 15:50:52.122601 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c-operator-scripts\") pod \"88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c\" (UID: \"88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c\") " Dec 09 15:50:52 crc kubenswrapper[4894]: I1209 15:50:52.123107 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3baa21af-058e-41cc-832b-6e6b138a33d2-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "3baa21af-058e-41cc-832b-6e6b138a33d2" (UID: "3baa21af-058e-41cc-832b-6e6b138a33d2"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:50:52 crc kubenswrapper[4894]: I1209 15:50:52.123138 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c" (UID: "88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:50:52 crc kubenswrapper[4894]: I1209 15:50:52.123838 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/3baa21af-058e-41cc-832b-6e6b138a33d2-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:52 crc kubenswrapper[4894]: I1209 15:50:52.123862 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:52 crc kubenswrapper[4894]: I1209 15:50:52.128047 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3baa21af-058e-41cc-832b-6e6b138a33d2-kube-api-access-j92dk" (OuterVolumeSpecName: "kube-api-access-j92dk") pod "3baa21af-058e-41cc-832b-6e6b138a33d2" (UID: "3baa21af-058e-41cc-832b-6e6b138a33d2"). InnerVolumeSpecName "kube-api-access-j92dk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:50:52 crc kubenswrapper[4894]: I1209 15:50:52.129023 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c-kube-api-access-qqvvt" (OuterVolumeSpecName: "kube-api-access-qqvvt") pod "88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c" (UID: "88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c"). InnerVolumeSpecName "kube-api-access-qqvvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:50:52 crc kubenswrapper[4894]: I1209 15:50:52.226013 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qqvvt\" (UniqueName: \"kubernetes.io/projected/88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c-kube-api-access-qqvvt\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:52 crc kubenswrapper[4894]: I1209 15:50:52.226046 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j92dk\" (UniqueName: \"kubernetes.io/projected/3baa21af-058e-41cc-832b-6e6b138a33d2-kube-api-access-j92dk\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:52 crc kubenswrapper[4894]: I1209 15:50:52.608539 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f44ed30-5746-4bc4-80b4-53b5622c13ae","Type":"ContainerStarted","Data":"850bd03ef9fa43c8f8c2ef7defdd5ef00c87d9a480d7de0b7c5bb70d6a413c96"} Dec 09 15:50:52 crc kubenswrapper[4894]: I1209 15:50:52.613081 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-5951-account-create-update-zl4pj" Dec 09 15:50:52 crc kubenswrapper[4894]: I1209 15:50:52.613078 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-5951-account-create-update-zl4pj" event={"ID":"3baa21af-058e-41cc-832b-6e6b138a33d2","Type":"ContainerDied","Data":"3d4cfdda7318b24caf98c0e9936c15803ae1af2ebf58e9e89a9dc23db0f3d1c1"} Dec 09 15:50:52 crc kubenswrapper[4894]: I1209 15:50:52.613160 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3d4cfdda7318b24caf98c0e9936c15803ae1af2ebf58e9e89a9dc23db0f3d1c1" Dec 09 15:50:52 crc kubenswrapper[4894]: I1209 15:50:52.625015 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-vt68w" event={"ID":"88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c","Type":"ContainerDied","Data":"8eadbc597a691bb8c6f1de7c942e49a967878dbf3416d1b7fb58f127709103c5"} Dec 09 15:50:52 crc kubenswrapper[4894]: I1209 15:50:52.625066 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8eadbc597a691bb8c6f1de7c942e49a967878dbf3416d1b7fb58f127709103c5" Dec 09 15:50:52 crc kubenswrapper[4894]: I1209 15:50:52.625030 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-vt68w" Dec 09 15:50:52 crc kubenswrapper[4894]: I1209 15:50:52.918037 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-5x62k" Dec 09 15:50:53 crc kubenswrapper[4894]: I1209 15:50:53.037546 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fd97k\" (UniqueName: \"kubernetes.io/projected/8e66b850-9ac9-43ce-977a-a8f61c80a2da-kube-api-access-fd97k\") pod \"8e66b850-9ac9-43ce-977a-a8f61c80a2da\" (UID: \"8e66b850-9ac9-43ce-977a-a8f61c80a2da\") " Dec 09 15:50:53 crc kubenswrapper[4894]: I1209 15:50:53.037630 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e66b850-9ac9-43ce-977a-a8f61c80a2da-operator-scripts\") pod \"8e66b850-9ac9-43ce-977a-a8f61c80a2da\" (UID: \"8e66b850-9ac9-43ce-977a-a8f61c80a2da\") " Dec 09 15:50:53 crc kubenswrapper[4894]: I1209 15:50:53.038030 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8e66b850-9ac9-43ce-977a-a8f61c80a2da-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "8e66b850-9ac9-43ce-977a-a8f61c80a2da" (UID: "8e66b850-9ac9-43ce-977a-a8f61c80a2da"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:50:53 crc kubenswrapper[4894]: I1209 15:50:53.038269 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/8e66b850-9ac9-43ce-977a-a8f61c80a2da-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:53 crc kubenswrapper[4894]: I1209 15:50:53.042793 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e66b850-9ac9-43ce-977a-a8f61c80a2da-kube-api-access-fd97k" (OuterVolumeSpecName: "kube-api-access-fd97k") pod "8e66b850-9ac9-43ce-977a-a8f61c80a2da" (UID: "8e66b850-9ac9-43ce-977a-a8f61c80a2da"). InnerVolumeSpecName "kube-api-access-fd97k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:50:53 crc kubenswrapper[4894]: I1209 15:50:53.140395 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fd97k\" (UniqueName: \"kubernetes.io/projected/8e66b850-9ac9-43ce-977a-a8f61c80a2da-kube-api-access-fd97k\") on node \"crc\" DevicePath \"\"" Dec 09 15:50:53 crc kubenswrapper[4894]: I1209 15:50:53.639081 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-5x62k" event={"ID":"8e66b850-9ac9-43ce-977a-a8f61c80a2da","Type":"ContainerDied","Data":"fe07168589d5548992764db66a4d1ed7fb2cd8ab123239643ecffeab01984502"} Dec 09 15:50:53 crc kubenswrapper[4894]: I1209 15:50:53.639416 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fe07168589d5548992764db66a4d1ed7fb2cd8ab123239643ecffeab01984502" Dec 09 15:50:53 crc kubenswrapper[4894]: I1209 15:50:53.639495 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-5x62k" Dec 09 15:50:59 crc kubenswrapper[4894]: I1209 15:50:59.701723 4894 generic.go:334] "Generic (PLEG): container finished" podID="22fdb6d9-4858-44bf-a856-04b66398260f" containerID="ebb2d23f3bfa50765824c94ae14d468a88a6a0b774c71abcf01a4eb038186069" exitCode=0 Dec 09 15:50:59 crc kubenswrapper[4894]: I1209 15:50:59.701807 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9b21-account-create-update-g5mj7" event={"ID":"22fdb6d9-4858-44bf-a856-04b66398260f","Type":"ContainerDied","Data":"ebb2d23f3bfa50765824c94ae14d468a88a6a0b774c71abcf01a4eb038186069"} Dec 09 15:50:59 crc kubenswrapper[4894]: I1209 15:50:59.704661 4894 generic.go:334] "Generic (PLEG): container finished" podID="423a2183-9395-4c3f-ab1c-1816c1d31fb9" containerID="a622ced4830d636677017d89b88e3a2170266e342db1cbe48a5537d668c423cd" exitCode=0 Dec 09 15:50:59 crc kubenswrapper[4894]: I1209 15:50:59.704701 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9b51-account-create-update-xvkxq" event={"ID":"423a2183-9395-4c3f-ab1c-1816c1d31fb9","Type":"ContainerDied","Data":"a622ced4830d636677017d89b88e3a2170266e342db1cbe48a5537d668c423cd"} Dec 09 15:51:00 crc kubenswrapper[4894]: I1209 15:51:00.715762 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f44ed30-5746-4bc4-80b4-53b5622c13ae","Type":"ContainerStarted","Data":"508625385a8336c7e3997d2fbde661793cd04cb4e82ce5e38ab427e18795d8a1"} Dec 09 15:51:00 crc kubenswrapper[4894]: I1209 15:51:00.722721 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f44ed30-5746-4bc4-80b4-53b5622c13ae","Type":"ContainerStarted","Data":"20be427d94fd19a5ffd3e00e2aaa74d24fd5e7e69f957c62b14dcf6f97e846aa"} Dec 09 15:51:01 crc kubenswrapper[4894]: I1209 15:51:01.104897 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9b51-account-create-update-xvkxq" Dec 09 15:51:01 crc kubenswrapper[4894]: I1209 15:51:01.119121 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9b21-account-create-update-g5mj7" Dec 09 15:51:01 crc kubenswrapper[4894]: I1209 15:51:01.190338 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22fdb6d9-4858-44bf-a856-04b66398260f-operator-scripts\") pod \"22fdb6d9-4858-44bf-a856-04b66398260f\" (UID: \"22fdb6d9-4858-44bf-a856-04b66398260f\") " Dec 09 15:51:01 crc kubenswrapper[4894]: I1209 15:51:01.190823 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mzdkj\" (UniqueName: \"kubernetes.io/projected/22fdb6d9-4858-44bf-a856-04b66398260f-kube-api-access-mzdkj\") pod \"22fdb6d9-4858-44bf-a856-04b66398260f\" (UID: \"22fdb6d9-4858-44bf-a856-04b66398260f\") " Dec 09 15:51:01 crc kubenswrapper[4894]: I1209 15:51:01.190880 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j79fm\" (UniqueName: \"kubernetes.io/projected/423a2183-9395-4c3f-ab1c-1816c1d31fb9-kube-api-access-j79fm\") pod \"423a2183-9395-4c3f-ab1c-1816c1d31fb9\" (UID: \"423a2183-9395-4c3f-ab1c-1816c1d31fb9\") " Dec 09 15:51:01 crc kubenswrapper[4894]: I1209 15:51:01.190915 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/423a2183-9395-4c3f-ab1c-1816c1d31fb9-operator-scripts\") pod \"423a2183-9395-4c3f-ab1c-1816c1d31fb9\" (UID: \"423a2183-9395-4c3f-ab1c-1816c1d31fb9\") " Dec 09 15:51:01 crc kubenswrapper[4894]: I1209 15:51:01.191489 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22fdb6d9-4858-44bf-a856-04b66398260f-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "22fdb6d9-4858-44bf-a856-04b66398260f" (UID: "22fdb6d9-4858-44bf-a856-04b66398260f"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:51:01 crc kubenswrapper[4894]: I1209 15:51:01.192824 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/423a2183-9395-4c3f-ab1c-1816c1d31fb9-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "423a2183-9395-4c3f-ab1c-1816c1d31fb9" (UID: "423a2183-9395-4c3f-ab1c-1816c1d31fb9"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:51:01 crc kubenswrapper[4894]: I1209 15:51:01.199461 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22fdb6d9-4858-44bf-a856-04b66398260f-kube-api-access-mzdkj" (OuterVolumeSpecName: "kube-api-access-mzdkj") pod "22fdb6d9-4858-44bf-a856-04b66398260f" (UID: "22fdb6d9-4858-44bf-a856-04b66398260f"). InnerVolumeSpecName "kube-api-access-mzdkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:51:01 crc kubenswrapper[4894]: I1209 15:51:01.199867 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/423a2183-9395-4c3f-ab1c-1816c1d31fb9-kube-api-access-j79fm" (OuterVolumeSpecName: "kube-api-access-j79fm") pod "423a2183-9395-4c3f-ab1c-1816c1d31fb9" (UID: "423a2183-9395-4c3f-ab1c-1816c1d31fb9"). InnerVolumeSpecName "kube-api-access-j79fm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:51:01 crc kubenswrapper[4894]: I1209 15:51:01.292451 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mzdkj\" (UniqueName: \"kubernetes.io/projected/22fdb6d9-4858-44bf-a856-04b66398260f-kube-api-access-mzdkj\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:01 crc kubenswrapper[4894]: I1209 15:51:01.292485 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j79fm\" (UniqueName: \"kubernetes.io/projected/423a2183-9395-4c3f-ab1c-1816c1d31fb9-kube-api-access-j79fm\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:01 crc kubenswrapper[4894]: I1209 15:51:01.292495 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/423a2183-9395-4c3f-ab1c-1816c1d31fb9-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:01 crc kubenswrapper[4894]: I1209 15:51:01.292504 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/22fdb6d9-4858-44bf-a856-04b66398260f-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:01 crc kubenswrapper[4894]: I1209 15:51:01.725152 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-9b51-account-create-update-xvkxq" event={"ID":"423a2183-9395-4c3f-ab1c-1816c1d31fb9","Type":"ContainerDied","Data":"dcf8f20731f814fbed4ffc86bb88fd6ca61da2e9240546c0720fbd6637e72817"} Dec 09 15:51:01 crc kubenswrapper[4894]: I1209 15:51:01.726088 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcf8f20731f814fbed4ffc86bb88fd6ca61da2e9240546c0720fbd6637e72817" Dec 09 15:51:01 crc kubenswrapper[4894]: I1209 15:51:01.726269 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-9b51-account-create-update-xvkxq" Dec 09 15:51:01 crc kubenswrapper[4894]: I1209 15:51:01.732053 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-9b21-account-create-update-g5mj7" event={"ID":"22fdb6d9-4858-44bf-a856-04b66398260f","Type":"ContainerDied","Data":"a92f04bf213953a71361f2ccf669889dd26cbbc7a9980cb6e831146a23937e3f"} Dec 09 15:51:01 crc kubenswrapper[4894]: I1209 15:51:01.732079 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a92f04bf213953a71361f2ccf669889dd26cbbc7a9980cb6e831146a23937e3f" Dec 09 15:51:01 crc kubenswrapper[4894]: I1209 15:51:01.732232 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-9b21-account-create-update-g5mj7" Dec 09 15:51:02 crc kubenswrapper[4894]: I1209 15:51:02.741526 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f44ed30-5746-4bc4-80b4-53b5622c13ae","Type":"ContainerStarted","Data":"e3f944728244560d7e24fb86911efb16f8f94af1bedc17c17020d9c9a22e88ab"} Dec 09 15:51:02 crc kubenswrapper[4894]: I1209 15:51:02.742027 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 15:51:03 crc kubenswrapper[4894]: I1209 15:51:03.794913 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.398908095 podStartE2EDuration="22.794890882s" podCreationTimestamp="2025-12-09 15:50:41 +0000 UTC" firstStartedPulling="2025-12-09 15:50:42.677036809 +0000 UTC m=+1136.996247478" lastFinishedPulling="2025-12-09 15:51:02.073019606 +0000 UTC m=+1156.392230265" observedRunningTime="2025-12-09 15:51:02.769032657 +0000 UTC m=+1157.088243326" watchObservedRunningTime="2025-12-09 15:51:03.794890882 +0000 UTC m=+1158.114101551" Dec 09 15:51:03 crc kubenswrapper[4894]: I1209 15:51:03.803965 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:51:04 crc kubenswrapper[4894]: I1209 15:51:04.758752 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4f44ed30-5746-4bc4-80b4-53b5622c13ae" containerName="ceilometer-central-agent" containerID="cri-o://850bd03ef9fa43c8f8c2ef7defdd5ef00c87d9a480d7de0b7c5bb70d6a413c96" gracePeriod=30 Dec 09 15:51:04 crc kubenswrapper[4894]: I1209 15:51:04.758777 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4f44ed30-5746-4bc4-80b4-53b5622c13ae" containerName="proxy-httpd" containerID="cri-o://e3f944728244560d7e24fb86911efb16f8f94af1bedc17c17020d9c9a22e88ab" gracePeriod=30 Dec 09 15:51:04 crc kubenswrapper[4894]: I1209 15:51:04.758808 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4f44ed30-5746-4bc4-80b4-53b5622c13ae" containerName="sg-core" containerID="cri-o://508625385a8336c7e3997d2fbde661793cd04cb4e82ce5e38ab427e18795d8a1" gracePeriod=30 Dec 09 15:51:04 crc kubenswrapper[4894]: I1209 15:51:04.758816 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4f44ed30-5746-4bc4-80b4-53b5622c13ae" containerName="ceilometer-notification-agent" containerID="cri-o://20be427d94fd19a5ffd3e00e2aaa74d24fd5e7e69f957c62b14dcf6f97e846aa" gracePeriod=30 Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.094759 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-b56ck"] Dec 09 15:51:05 crc kubenswrapper[4894]: E1209 15:51:05.095807 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c" containerName="mariadb-database-create" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.095903 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c" containerName="mariadb-database-create" Dec 09 15:51:05 crc kubenswrapper[4894]: E1209 15:51:05.095964 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22fdb6d9-4858-44bf-a856-04b66398260f" containerName="mariadb-account-create-update" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.096032 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="22fdb6d9-4858-44bf-a856-04b66398260f" containerName="mariadb-account-create-update" Dec 09 15:51:05 crc kubenswrapper[4894]: E1209 15:51:05.096097 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e66b850-9ac9-43ce-977a-a8f61c80a2da" containerName="mariadb-database-create" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.096154 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e66b850-9ac9-43ce-977a-a8f61c80a2da" containerName="mariadb-database-create" Dec 09 15:51:05 crc kubenswrapper[4894]: E1209 15:51:05.096215 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="423a2183-9395-4c3f-ab1c-1816c1d31fb9" containerName="mariadb-account-create-update" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.096506 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="423a2183-9395-4c3f-ab1c-1816c1d31fb9" containerName="mariadb-account-create-update" Dec 09 15:51:05 crc kubenswrapper[4894]: E1209 15:51:05.096574 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3baa21af-058e-41cc-832b-6e6b138a33d2" containerName="mariadb-account-create-update" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.096583 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="3baa21af-058e-41cc-832b-6e6b138a33d2" containerName="mariadb-account-create-update" Dec 09 15:51:05 crc kubenswrapper[4894]: E1209 15:51:05.096604 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea9963b9-ff7e-4d92-9cf1-7a5108a33b76" containerName="mariadb-database-create" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.096610 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea9963b9-ff7e-4d92-9cf1-7a5108a33b76" containerName="mariadb-database-create" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.096937 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="423a2183-9395-4c3f-ab1c-1816c1d31fb9" containerName="mariadb-account-create-update" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.096956 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c" containerName="mariadb-database-create" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.096966 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="3baa21af-058e-41cc-832b-6e6b138a33d2" containerName="mariadb-account-create-update" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.096988 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e66b850-9ac9-43ce-977a-a8f61c80a2da" containerName="mariadb-database-create" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.096998 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea9963b9-ff7e-4d92-9cf1-7a5108a33b76" containerName="mariadb-database-create" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.097006 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="22fdb6d9-4858-44bf-a856-04b66398260f" containerName="mariadb-account-create-update" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.097600 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-b56ck" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.100426 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.100661 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-6fv64" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.100726 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.107943 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-b56ck"] Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.159004 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ed12da4-c785-4d59-a3d7-c485b97013bc-config-data\") pod \"nova-cell0-conductor-db-sync-b56ck\" (UID: \"8ed12da4-c785-4d59-a3d7-c485b97013bc\") " pod="openstack/nova-cell0-conductor-db-sync-b56ck" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.159159 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ed12da4-c785-4d59-a3d7-c485b97013bc-scripts\") pod \"nova-cell0-conductor-db-sync-b56ck\" (UID: \"8ed12da4-c785-4d59-a3d7-c485b97013bc\") " pod="openstack/nova-cell0-conductor-db-sync-b56ck" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.159182 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gzhv\" (UniqueName: \"kubernetes.io/projected/8ed12da4-c785-4d59-a3d7-c485b97013bc-kube-api-access-5gzhv\") pod \"nova-cell0-conductor-db-sync-b56ck\" (UID: \"8ed12da4-c785-4d59-a3d7-c485b97013bc\") " pod="openstack/nova-cell0-conductor-db-sync-b56ck" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.159204 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ed12da4-c785-4d59-a3d7-c485b97013bc-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-b56ck\" (UID: \"8ed12da4-c785-4d59-a3d7-c485b97013bc\") " pod="openstack/nova-cell0-conductor-db-sync-b56ck" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.261134 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ed12da4-c785-4d59-a3d7-c485b97013bc-scripts\") pod \"nova-cell0-conductor-db-sync-b56ck\" (UID: \"8ed12da4-c785-4d59-a3d7-c485b97013bc\") " pod="openstack/nova-cell0-conductor-db-sync-b56ck" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.261183 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gzhv\" (UniqueName: \"kubernetes.io/projected/8ed12da4-c785-4d59-a3d7-c485b97013bc-kube-api-access-5gzhv\") pod \"nova-cell0-conductor-db-sync-b56ck\" (UID: \"8ed12da4-c785-4d59-a3d7-c485b97013bc\") " pod="openstack/nova-cell0-conductor-db-sync-b56ck" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.261204 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ed12da4-c785-4d59-a3d7-c485b97013bc-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-b56ck\" (UID: \"8ed12da4-c785-4d59-a3d7-c485b97013bc\") " pod="openstack/nova-cell0-conductor-db-sync-b56ck" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.261240 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ed12da4-c785-4d59-a3d7-c485b97013bc-config-data\") pod \"nova-cell0-conductor-db-sync-b56ck\" (UID: \"8ed12da4-c785-4d59-a3d7-c485b97013bc\") " pod="openstack/nova-cell0-conductor-db-sync-b56ck" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.266764 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ed12da4-c785-4d59-a3d7-c485b97013bc-scripts\") pod \"nova-cell0-conductor-db-sync-b56ck\" (UID: \"8ed12da4-c785-4d59-a3d7-c485b97013bc\") " pod="openstack/nova-cell0-conductor-db-sync-b56ck" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.268501 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ed12da4-c785-4d59-a3d7-c485b97013bc-config-data\") pod \"nova-cell0-conductor-db-sync-b56ck\" (UID: \"8ed12da4-c785-4d59-a3d7-c485b97013bc\") " pod="openstack/nova-cell0-conductor-db-sync-b56ck" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.269144 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ed12da4-c785-4d59-a3d7-c485b97013bc-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-b56ck\" (UID: \"8ed12da4-c785-4d59-a3d7-c485b97013bc\") " pod="openstack/nova-cell0-conductor-db-sync-b56ck" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.284237 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gzhv\" (UniqueName: \"kubernetes.io/projected/8ed12da4-c785-4d59-a3d7-c485b97013bc-kube-api-access-5gzhv\") pod \"nova-cell0-conductor-db-sync-b56ck\" (UID: \"8ed12da4-c785-4d59-a3d7-c485b97013bc\") " pod="openstack/nova-cell0-conductor-db-sync-b56ck" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.450966 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-b56ck" Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.782951 4894 generic.go:334] "Generic (PLEG): container finished" podID="4f44ed30-5746-4bc4-80b4-53b5622c13ae" containerID="e3f944728244560d7e24fb86911efb16f8f94af1bedc17c17020d9c9a22e88ab" exitCode=0 Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.783200 4894 generic.go:334] "Generic (PLEG): container finished" podID="4f44ed30-5746-4bc4-80b4-53b5622c13ae" containerID="508625385a8336c7e3997d2fbde661793cd04cb4e82ce5e38ab427e18795d8a1" exitCode=2 Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.782998 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f44ed30-5746-4bc4-80b4-53b5622c13ae","Type":"ContainerDied","Data":"e3f944728244560d7e24fb86911efb16f8f94af1bedc17c17020d9c9a22e88ab"} Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.783228 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f44ed30-5746-4bc4-80b4-53b5622c13ae","Type":"ContainerDied","Data":"508625385a8336c7e3997d2fbde661793cd04cb4e82ce5e38ab427e18795d8a1"} Dec 09 15:51:05 crc kubenswrapper[4894]: W1209 15:51:05.910182 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8ed12da4_c785_4d59_a3d7_c485b97013bc.slice/crio-a8485224bf9df9c87c412470d36c1071122e00b04afd66a918c84c4927d70dc4 WatchSource:0}: Error finding container a8485224bf9df9c87c412470d36c1071122e00b04afd66a918c84c4927d70dc4: Status 404 returned error can't find the container with id a8485224bf9df9c87c412470d36c1071122e00b04afd66a918c84c4927d70dc4 Dec 09 15:51:05 crc kubenswrapper[4894]: I1209 15:51:05.911012 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-b56ck"] Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.636369 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.691390 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-scripts\") pod \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.691787 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f44ed30-5746-4bc4-80b4-53b5622c13ae-run-httpd\") pod \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.691896 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-combined-ca-bundle\") pod \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.691981 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-config-data\") pod \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.692129 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f44ed30-5746-4bc4-80b4-53b5622c13ae-log-httpd\") pod \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.692163 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbwsd\" (UniqueName: \"kubernetes.io/projected/4f44ed30-5746-4bc4-80b4-53b5622c13ae-kube-api-access-rbwsd\") pod \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.692380 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-sg-core-conf-yaml\") pod \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\" (UID: \"4f44ed30-5746-4bc4-80b4-53b5622c13ae\") " Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.692621 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f44ed30-5746-4bc4-80b4-53b5622c13ae-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4f44ed30-5746-4bc4-80b4-53b5622c13ae" (UID: "4f44ed30-5746-4bc4-80b4-53b5622c13ae"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.692661 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4f44ed30-5746-4bc4-80b4-53b5622c13ae-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4f44ed30-5746-4bc4-80b4-53b5622c13ae" (UID: "4f44ed30-5746-4bc4-80b4-53b5622c13ae"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.693348 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f44ed30-5746-4bc4-80b4-53b5622c13ae-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.693374 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4f44ed30-5746-4bc4-80b4-53b5622c13ae-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.700280 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-scripts" (OuterVolumeSpecName: "scripts") pod "4f44ed30-5746-4bc4-80b4-53b5622c13ae" (UID: "4f44ed30-5746-4bc4-80b4-53b5622c13ae"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.700370 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4f44ed30-5746-4bc4-80b4-53b5622c13ae-kube-api-access-rbwsd" (OuterVolumeSpecName: "kube-api-access-rbwsd") pod "4f44ed30-5746-4bc4-80b4-53b5622c13ae" (UID: "4f44ed30-5746-4bc4-80b4-53b5622c13ae"). InnerVolumeSpecName "kube-api-access-rbwsd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.722407 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4f44ed30-5746-4bc4-80b4-53b5622c13ae" (UID: "4f44ed30-5746-4bc4-80b4-53b5622c13ae"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.794659 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbwsd\" (UniqueName: \"kubernetes.io/projected/4f44ed30-5746-4bc4-80b4-53b5622c13ae-kube-api-access-rbwsd\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.795019 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.795097 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.797537 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-b56ck" event={"ID":"8ed12da4-c785-4d59-a3d7-c485b97013bc","Type":"ContainerStarted","Data":"a8485224bf9df9c87c412470d36c1071122e00b04afd66a918c84c4927d70dc4"} Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.803719 4894 generic.go:334] "Generic (PLEG): container finished" podID="4f44ed30-5746-4bc4-80b4-53b5622c13ae" containerID="20be427d94fd19a5ffd3e00e2aaa74d24fd5e7e69f957c62b14dcf6f97e846aa" exitCode=0 Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.803901 4894 generic.go:334] "Generic (PLEG): container finished" podID="4f44ed30-5746-4bc4-80b4-53b5622c13ae" containerID="850bd03ef9fa43c8f8c2ef7defdd5ef00c87d9a480d7de0b7c5bb70d6a413c96" exitCode=0 Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.803930 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f44ed30-5746-4bc4-80b4-53b5622c13ae","Type":"ContainerDied","Data":"20be427d94fd19a5ffd3e00e2aaa74d24fd5e7e69f957c62b14dcf6f97e846aa"} Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.804253 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f44ed30-5746-4bc4-80b4-53b5622c13ae","Type":"ContainerDied","Data":"850bd03ef9fa43c8f8c2ef7defdd5ef00c87d9a480d7de0b7c5bb70d6a413c96"} Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.804332 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4f44ed30-5746-4bc4-80b4-53b5622c13ae","Type":"ContainerDied","Data":"156a48ba07176f332f47253915b8f7d179afeec2b7a8db2fcbb0bce90e27bd83"} Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.804424 4894 scope.go:117] "RemoveContainer" containerID="e3f944728244560d7e24fb86911efb16f8f94af1bedc17c17020d9c9a22e88ab" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.804026 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.806283 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4f44ed30-5746-4bc4-80b4-53b5622c13ae" (UID: "4f44ed30-5746-4bc4-80b4-53b5622c13ae"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.828919 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-config-data" (OuterVolumeSpecName: "config-data") pod "4f44ed30-5746-4bc4-80b4-53b5622c13ae" (UID: "4f44ed30-5746-4bc4-80b4-53b5622c13ae"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.838917 4894 scope.go:117] "RemoveContainer" containerID="508625385a8336c7e3997d2fbde661793cd04cb4e82ce5e38ab427e18795d8a1" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.879093 4894 scope.go:117] "RemoveContainer" containerID="20be427d94fd19a5ffd3e00e2aaa74d24fd5e7e69f957c62b14dcf6f97e846aa" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.896907 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.896949 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4f44ed30-5746-4bc4-80b4-53b5622c13ae-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.950975 4894 scope.go:117] "RemoveContainer" containerID="850bd03ef9fa43c8f8c2ef7defdd5ef00c87d9a480d7de0b7c5bb70d6a413c96" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.976261 4894 scope.go:117] "RemoveContainer" containerID="e3f944728244560d7e24fb86911efb16f8f94af1bedc17c17020d9c9a22e88ab" Dec 09 15:51:06 crc kubenswrapper[4894]: E1209 15:51:06.976884 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3f944728244560d7e24fb86911efb16f8f94af1bedc17c17020d9c9a22e88ab\": container with ID starting with e3f944728244560d7e24fb86911efb16f8f94af1bedc17c17020d9c9a22e88ab not found: ID does not exist" containerID="e3f944728244560d7e24fb86911efb16f8f94af1bedc17c17020d9c9a22e88ab" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.976930 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3f944728244560d7e24fb86911efb16f8f94af1bedc17c17020d9c9a22e88ab"} err="failed to get container status \"e3f944728244560d7e24fb86911efb16f8f94af1bedc17c17020d9c9a22e88ab\": rpc error: code = NotFound desc = could not find container \"e3f944728244560d7e24fb86911efb16f8f94af1bedc17c17020d9c9a22e88ab\": container with ID starting with e3f944728244560d7e24fb86911efb16f8f94af1bedc17c17020d9c9a22e88ab not found: ID does not exist" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.976962 4894 scope.go:117] "RemoveContainer" containerID="508625385a8336c7e3997d2fbde661793cd04cb4e82ce5e38ab427e18795d8a1" Dec 09 15:51:06 crc kubenswrapper[4894]: E1209 15:51:06.977251 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"508625385a8336c7e3997d2fbde661793cd04cb4e82ce5e38ab427e18795d8a1\": container with ID starting with 508625385a8336c7e3997d2fbde661793cd04cb4e82ce5e38ab427e18795d8a1 not found: ID does not exist" containerID="508625385a8336c7e3997d2fbde661793cd04cb4e82ce5e38ab427e18795d8a1" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.977296 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"508625385a8336c7e3997d2fbde661793cd04cb4e82ce5e38ab427e18795d8a1"} err="failed to get container status \"508625385a8336c7e3997d2fbde661793cd04cb4e82ce5e38ab427e18795d8a1\": rpc error: code = NotFound desc = could not find container \"508625385a8336c7e3997d2fbde661793cd04cb4e82ce5e38ab427e18795d8a1\": container with ID starting with 508625385a8336c7e3997d2fbde661793cd04cb4e82ce5e38ab427e18795d8a1 not found: ID does not exist" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.977322 4894 scope.go:117] "RemoveContainer" containerID="20be427d94fd19a5ffd3e00e2aaa74d24fd5e7e69f957c62b14dcf6f97e846aa" Dec 09 15:51:06 crc kubenswrapper[4894]: E1209 15:51:06.977616 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20be427d94fd19a5ffd3e00e2aaa74d24fd5e7e69f957c62b14dcf6f97e846aa\": container with ID starting with 20be427d94fd19a5ffd3e00e2aaa74d24fd5e7e69f957c62b14dcf6f97e846aa not found: ID does not exist" containerID="20be427d94fd19a5ffd3e00e2aaa74d24fd5e7e69f957c62b14dcf6f97e846aa" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.977672 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20be427d94fd19a5ffd3e00e2aaa74d24fd5e7e69f957c62b14dcf6f97e846aa"} err="failed to get container status \"20be427d94fd19a5ffd3e00e2aaa74d24fd5e7e69f957c62b14dcf6f97e846aa\": rpc error: code = NotFound desc = could not find container \"20be427d94fd19a5ffd3e00e2aaa74d24fd5e7e69f957c62b14dcf6f97e846aa\": container with ID starting with 20be427d94fd19a5ffd3e00e2aaa74d24fd5e7e69f957c62b14dcf6f97e846aa not found: ID does not exist" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.977693 4894 scope.go:117] "RemoveContainer" containerID="850bd03ef9fa43c8f8c2ef7defdd5ef00c87d9a480d7de0b7c5bb70d6a413c96" Dec 09 15:51:06 crc kubenswrapper[4894]: E1209 15:51:06.977931 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"850bd03ef9fa43c8f8c2ef7defdd5ef00c87d9a480d7de0b7c5bb70d6a413c96\": container with ID starting with 850bd03ef9fa43c8f8c2ef7defdd5ef00c87d9a480d7de0b7c5bb70d6a413c96 not found: ID does not exist" containerID="850bd03ef9fa43c8f8c2ef7defdd5ef00c87d9a480d7de0b7c5bb70d6a413c96" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.977959 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"850bd03ef9fa43c8f8c2ef7defdd5ef00c87d9a480d7de0b7c5bb70d6a413c96"} err="failed to get container status \"850bd03ef9fa43c8f8c2ef7defdd5ef00c87d9a480d7de0b7c5bb70d6a413c96\": rpc error: code = NotFound desc = could not find container \"850bd03ef9fa43c8f8c2ef7defdd5ef00c87d9a480d7de0b7c5bb70d6a413c96\": container with ID starting with 850bd03ef9fa43c8f8c2ef7defdd5ef00c87d9a480d7de0b7c5bb70d6a413c96 not found: ID does not exist" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.977980 4894 scope.go:117] "RemoveContainer" containerID="e3f944728244560d7e24fb86911efb16f8f94af1bedc17c17020d9c9a22e88ab" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.978235 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3f944728244560d7e24fb86911efb16f8f94af1bedc17c17020d9c9a22e88ab"} err="failed to get container status \"e3f944728244560d7e24fb86911efb16f8f94af1bedc17c17020d9c9a22e88ab\": rpc error: code = NotFound desc = could not find container \"e3f944728244560d7e24fb86911efb16f8f94af1bedc17c17020d9c9a22e88ab\": container with ID starting with e3f944728244560d7e24fb86911efb16f8f94af1bedc17c17020d9c9a22e88ab not found: ID does not exist" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.978256 4894 scope.go:117] "RemoveContainer" containerID="508625385a8336c7e3997d2fbde661793cd04cb4e82ce5e38ab427e18795d8a1" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.978774 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"508625385a8336c7e3997d2fbde661793cd04cb4e82ce5e38ab427e18795d8a1"} err="failed to get container status \"508625385a8336c7e3997d2fbde661793cd04cb4e82ce5e38ab427e18795d8a1\": rpc error: code = NotFound desc = could not find container \"508625385a8336c7e3997d2fbde661793cd04cb4e82ce5e38ab427e18795d8a1\": container with ID starting with 508625385a8336c7e3997d2fbde661793cd04cb4e82ce5e38ab427e18795d8a1 not found: ID does not exist" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.978801 4894 scope.go:117] "RemoveContainer" containerID="20be427d94fd19a5ffd3e00e2aaa74d24fd5e7e69f957c62b14dcf6f97e846aa" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.979103 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20be427d94fd19a5ffd3e00e2aaa74d24fd5e7e69f957c62b14dcf6f97e846aa"} err="failed to get container status \"20be427d94fd19a5ffd3e00e2aaa74d24fd5e7e69f957c62b14dcf6f97e846aa\": rpc error: code = NotFound desc = could not find container \"20be427d94fd19a5ffd3e00e2aaa74d24fd5e7e69f957c62b14dcf6f97e846aa\": container with ID starting with 20be427d94fd19a5ffd3e00e2aaa74d24fd5e7e69f957c62b14dcf6f97e846aa not found: ID does not exist" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.979126 4894 scope.go:117] "RemoveContainer" containerID="850bd03ef9fa43c8f8c2ef7defdd5ef00c87d9a480d7de0b7c5bb70d6a413c96" Dec 09 15:51:06 crc kubenswrapper[4894]: I1209 15:51:06.979495 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"850bd03ef9fa43c8f8c2ef7defdd5ef00c87d9a480d7de0b7c5bb70d6a413c96"} err="failed to get container status \"850bd03ef9fa43c8f8c2ef7defdd5ef00c87d9a480d7de0b7c5bb70d6a413c96\": rpc error: code = NotFound desc = could not find container \"850bd03ef9fa43c8f8c2ef7defdd5ef00c87d9a480d7de0b7c5bb70d6a413c96\": container with ID starting with 850bd03ef9fa43c8f8c2ef7defdd5ef00c87d9a480d7de0b7c5bb70d6a413c96 not found: ID does not exist" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.144665 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.158587 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.173313 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:51:07 crc kubenswrapper[4894]: E1209 15:51:07.177683 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f44ed30-5746-4bc4-80b4-53b5622c13ae" containerName="sg-core" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.177727 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f44ed30-5746-4bc4-80b4-53b5622c13ae" containerName="sg-core" Dec 09 15:51:07 crc kubenswrapper[4894]: E1209 15:51:07.177787 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f44ed30-5746-4bc4-80b4-53b5622c13ae" containerName="ceilometer-central-agent" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.177796 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f44ed30-5746-4bc4-80b4-53b5622c13ae" containerName="ceilometer-central-agent" Dec 09 15:51:07 crc kubenswrapper[4894]: E1209 15:51:07.177829 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f44ed30-5746-4bc4-80b4-53b5622c13ae" containerName="proxy-httpd" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.177837 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f44ed30-5746-4bc4-80b4-53b5622c13ae" containerName="proxy-httpd" Dec 09 15:51:07 crc kubenswrapper[4894]: E1209 15:51:07.177874 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4f44ed30-5746-4bc4-80b4-53b5622c13ae" containerName="ceilometer-notification-agent" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.177885 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="4f44ed30-5746-4bc4-80b4-53b5622c13ae" containerName="ceilometer-notification-agent" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.178303 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f44ed30-5746-4bc4-80b4-53b5622c13ae" containerName="ceilometer-central-agent" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.178337 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f44ed30-5746-4bc4-80b4-53b5622c13ae" containerName="ceilometer-notification-agent" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.178347 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f44ed30-5746-4bc4-80b4-53b5622c13ae" containerName="sg-core" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.178370 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="4f44ed30-5746-4bc4-80b4-53b5622c13ae" containerName="proxy-httpd" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.190083 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.204278 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.204705 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.221423 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.305015 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.305081 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.305107 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-scripts\") pod \"ceilometer-0\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.305136 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-config-data\") pod \"ceilometer-0\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.305201 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/992661f7-3ddd-40fb-9790-268cc5c0b83f-log-httpd\") pod \"ceilometer-0\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.305220 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdrh7\" (UniqueName: \"kubernetes.io/projected/992661f7-3ddd-40fb-9790-268cc5c0b83f-kube-api-access-cdrh7\") pod \"ceilometer-0\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.305245 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/992661f7-3ddd-40fb-9790-268cc5c0b83f-run-httpd\") pod \"ceilometer-0\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.407316 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/992661f7-3ddd-40fb-9790-268cc5c0b83f-log-httpd\") pod \"ceilometer-0\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.407574 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdrh7\" (UniqueName: \"kubernetes.io/projected/992661f7-3ddd-40fb-9790-268cc5c0b83f-kube-api-access-cdrh7\") pod \"ceilometer-0\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.407620 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/992661f7-3ddd-40fb-9790-268cc5c0b83f-run-httpd\") pod \"ceilometer-0\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.407733 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.407757 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.407778 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-scripts\") pod \"ceilometer-0\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.407804 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-config-data\") pod \"ceilometer-0\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.409095 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/992661f7-3ddd-40fb-9790-268cc5c0b83f-run-httpd\") pod \"ceilometer-0\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.409397 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/992661f7-3ddd-40fb-9790-268cc5c0b83f-log-httpd\") pod \"ceilometer-0\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.429289 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.430644 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.431703 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-config-data\") pod \"ceilometer-0\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.435494 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-scripts\") pod \"ceilometer-0\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.460527 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdrh7\" (UniqueName: \"kubernetes.io/projected/992661f7-3ddd-40fb-9790-268cc5c0b83f-kube-api-access-cdrh7\") pod \"ceilometer-0\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.530913 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:51:07 crc kubenswrapper[4894]: I1209 15:51:07.996931 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:51:08 crc kubenswrapper[4894]: I1209 15:51:08.119140 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4f44ed30-5746-4bc4-80b4-53b5622c13ae" path="/var/lib/kubelet/pods/4f44ed30-5746-4bc4-80b4-53b5622c13ae/volumes" Dec 09 15:51:08 crc kubenswrapper[4894]: I1209 15:51:08.832565 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"992661f7-3ddd-40fb-9790-268cc5c0b83f","Type":"ContainerStarted","Data":"e84c567e253439ba55fc26e0927f002286137edfcde42c4d400f2abb7345a1b6"} Dec 09 15:51:13 crc kubenswrapper[4894]: I1209 15:51:13.788559 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:51:16 crc kubenswrapper[4894]: I1209 15:51:16.905196 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-b56ck" event={"ID":"8ed12da4-c785-4d59-a3d7-c485b97013bc","Type":"ContainerStarted","Data":"eb0e84f294bc81396d21a68308cbb6daf5f04026d580ce87f6bd2297b9c4fb4e"} Dec 09 15:51:16 crc kubenswrapper[4894]: I1209 15:51:16.909435 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"992661f7-3ddd-40fb-9790-268cc5c0b83f","Type":"ContainerStarted","Data":"7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395"} Dec 09 15:51:16 crc kubenswrapper[4894]: I1209 15:51:16.926140 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-b56ck" podStartSLOduration=1.991501003 podStartE2EDuration="11.926124138s" podCreationTimestamp="2025-12-09 15:51:05 +0000 UTC" firstStartedPulling="2025-12-09 15:51:05.912566924 +0000 UTC m=+1160.231777593" lastFinishedPulling="2025-12-09 15:51:15.847190059 +0000 UTC m=+1170.166400728" observedRunningTime="2025-12-09 15:51:16.917731476 +0000 UTC m=+1171.236942165" watchObservedRunningTime="2025-12-09 15:51:16.926124138 +0000 UTC m=+1171.245334807" Dec 09 15:51:23 crc kubenswrapper[4894]: I1209 15:51:23.978308 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"992661f7-3ddd-40fb-9790-268cc5c0b83f","Type":"ContainerStarted","Data":"99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3"} Dec 09 15:51:25 crc kubenswrapper[4894]: I1209 15:51:25.995397 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"992661f7-3ddd-40fb-9790-268cc5c0b83f","Type":"ContainerStarted","Data":"216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6"} Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.007474 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"992661f7-3ddd-40fb-9790-268cc5c0b83f","Type":"ContainerStarted","Data":"e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429"} Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.007855 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.007747 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="992661f7-3ddd-40fb-9790-268cc5c0b83f" containerName="ceilometer-central-agent" containerID="cri-o://7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395" gracePeriod=30 Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.007938 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="992661f7-3ddd-40fb-9790-268cc5c0b83f" containerName="sg-core" containerID="cri-o://216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6" gracePeriod=30 Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.007962 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="992661f7-3ddd-40fb-9790-268cc5c0b83f" containerName="ceilometer-notification-agent" containerID="cri-o://99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3" gracePeriod=30 Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.007979 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="992661f7-3ddd-40fb-9790-268cc5c0b83f" containerName="proxy-httpd" containerID="cri-o://e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429" gracePeriod=30 Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.051487 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.547233926 podStartE2EDuration="20.051467779s" podCreationTimestamp="2025-12-09 15:51:07 +0000 UTC" firstStartedPulling="2025-12-09 15:51:08.026753987 +0000 UTC m=+1162.345964656" lastFinishedPulling="2025-12-09 15:51:26.5309878 +0000 UTC m=+1180.850198509" observedRunningTime="2025-12-09 15:51:27.045091487 +0000 UTC m=+1181.364302156" watchObservedRunningTime="2025-12-09 15:51:27.051467779 +0000 UTC m=+1181.370678448" Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.686009 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.826008 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-config-data\") pod \"992661f7-3ddd-40fb-9790-268cc5c0b83f\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.826080 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdrh7\" (UniqueName: \"kubernetes.io/projected/992661f7-3ddd-40fb-9790-268cc5c0b83f-kube-api-access-cdrh7\") pod \"992661f7-3ddd-40fb-9790-268cc5c0b83f\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.826124 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/992661f7-3ddd-40fb-9790-268cc5c0b83f-run-httpd\") pod \"992661f7-3ddd-40fb-9790-268cc5c0b83f\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.826229 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/992661f7-3ddd-40fb-9790-268cc5c0b83f-log-httpd\") pod \"992661f7-3ddd-40fb-9790-268cc5c0b83f\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.826332 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-sg-core-conf-yaml\") pod \"992661f7-3ddd-40fb-9790-268cc5c0b83f\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.826434 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-scripts\") pod \"992661f7-3ddd-40fb-9790-268cc5c0b83f\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.826486 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-combined-ca-bundle\") pod \"992661f7-3ddd-40fb-9790-268cc5c0b83f\" (UID: \"992661f7-3ddd-40fb-9790-268cc5c0b83f\") " Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.827277 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/992661f7-3ddd-40fb-9790-268cc5c0b83f-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "992661f7-3ddd-40fb-9790-268cc5c0b83f" (UID: "992661f7-3ddd-40fb-9790-268cc5c0b83f"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.827311 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/992661f7-3ddd-40fb-9790-268cc5c0b83f-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "992661f7-3ddd-40fb-9790-268cc5c0b83f" (UID: "992661f7-3ddd-40fb-9790-268cc5c0b83f"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.833501 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-scripts" (OuterVolumeSpecName: "scripts") pod "992661f7-3ddd-40fb-9790-268cc5c0b83f" (UID: "992661f7-3ddd-40fb-9790-268cc5c0b83f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.834120 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/992661f7-3ddd-40fb-9790-268cc5c0b83f-kube-api-access-cdrh7" (OuterVolumeSpecName: "kube-api-access-cdrh7") pod "992661f7-3ddd-40fb-9790-268cc5c0b83f" (UID: "992661f7-3ddd-40fb-9790-268cc5c0b83f"). InnerVolumeSpecName "kube-api-access-cdrh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.880562 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "992661f7-3ddd-40fb-9790-268cc5c0b83f" (UID: "992661f7-3ddd-40fb-9790-268cc5c0b83f"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.928555 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.928584 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdrh7\" (UniqueName: \"kubernetes.io/projected/992661f7-3ddd-40fb-9790-268cc5c0b83f-kube-api-access-cdrh7\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.928593 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/992661f7-3ddd-40fb-9790-268cc5c0b83f-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.928601 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/992661f7-3ddd-40fb-9790-268cc5c0b83f-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.928609 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.952920 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-config-data" (OuterVolumeSpecName: "config-data") pod "992661f7-3ddd-40fb-9790-268cc5c0b83f" (UID: "992661f7-3ddd-40fb-9790-268cc5c0b83f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:51:27 crc kubenswrapper[4894]: I1209 15:51:27.954194 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "992661f7-3ddd-40fb-9790-268cc5c0b83f" (UID: "992661f7-3ddd-40fb-9790-268cc5c0b83f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.030596 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.030704 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/992661f7-3ddd-40fb-9790-268cc5c0b83f-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.031550 4894 generic.go:334] "Generic (PLEG): container finished" podID="992661f7-3ddd-40fb-9790-268cc5c0b83f" containerID="e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429" exitCode=0 Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.031599 4894 generic.go:334] "Generic (PLEG): container finished" podID="992661f7-3ddd-40fb-9790-268cc5c0b83f" containerID="216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6" exitCode=2 Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.031615 4894 generic.go:334] "Generic (PLEG): container finished" podID="992661f7-3ddd-40fb-9790-268cc5c0b83f" containerID="99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3" exitCode=0 Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.031634 4894 generic.go:334] "Generic (PLEG): container finished" podID="992661f7-3ddd-40fb-9790-268cc5c0b83f" containerID="7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395" exitCode=0 Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.031702 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"992661f7-3ddd-40fb-9790-268cc5c0b83f","Type":"ContainerDied","Data":"e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429"} Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.031748 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"992661f7-3ddd-40fb-9790-268cc5c0b83f","Type":"ContainerDied","Data":"216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6"} Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.031770 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"992661f7-3ddd-40fb-9790-268cc5c0b83f","Type":"ContainerDied","Data":"99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3"} Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.031788 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"992661f7-3ddd-40fb-9790-268cc5c0b83f","Type":"ContainerDied","Data":"7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395"} Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.031817 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"992661f7-3ddd-40fb-9790-268cc5c0b83f","Type":"ContainerDied","Data":"e84c567e253439ba55fc26e0927f002286137edfcde42c4d400f2abb7345a1b6"} Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.031844 4894 scope.go:117] "RemoveContainer" containerID="e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.033374 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.052447 4894 scope.go:117] "RemoveContainer" containerID="216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.083517 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.101608 4894 scope.go:117] "RemoveContainer" containerID="99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.120708 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.122259 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:51:28 crc kubenswrapper[4894]: E1209 15:51:28.122796 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="992661f7-3ddd-40fb-9790-268cc5c0b83f" containerName="ceilometer-notification-agent" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.122820 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="992661f7-3ddd-40fb-9790-268cc5c0b83f" containerName="ceilometer-notification-agent" Dec 09 15:51:28 crc kubenswrapper[4894]: E1209 15:51:28.122844 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="992661f7-3ddd-40fb-9790-268cc5c0b83f" containerName="ceilometer-central-agent" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.122853 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="992661f7-3ddd-40fb-9790-268cc5c0b83f" containerName="ceilometer-central-agent" Dec 09 15:51:28 crc kubenswrapper[4894]: E1209 15:51:28.122866 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="992661f7-3ddd-40fb-9790-268cc5c0b83f" containerName="sg-core" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.122874 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="992661f7-3ddd-40fb-9790-268cc5c0b83f" containerName="sg-core" Dec 09 15:51:28 crc kubenswrapper[4894]: E1209 15:51:28.122894 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="992661f7-3ddd-40fb-9790-268cc5c0b83f" containerName="proxy-httpd" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.122901 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="992661f7-3ddd-40fb-9790-268cc5c0b83f" containerName="proxy-httpd" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.123121 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="992661f7-3ddd-40fb-9790-268cc5c0b83f" containerName="sg-core" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.123139 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="992661f7-3ddd-40fb-9790-268cc5c0b83f" containerName="ceilometer-central-agent" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.123152 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="992661f7-3ddd-40fb-9790-268cc5c0b83f" containerName="proxy-httpd" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.123168 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="992661f7-3ddd-40fb-9790-268cc5c0b83f" containerName="ceilometer-notification-agent" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.129103 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.131823 4894 scope.go:117] "RemoveContainer" containerID="7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.132701 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.133419 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.133485 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.165197 4894 scope.go:117] "RemoveContainer" containerID="e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429" Dec 09 15:51:28 crc kubenswrapper[4894]: E1209 15:51:28.165983 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429\": container with ID starting with e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429 not found: ID does not exist" containerID="e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.166022 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429"} err="failed to get container status \"e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429\": rpc error: code = NotFound desc = could not find container \"e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429\": container with ID starting with e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429 not found: ID does not exist" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.166047 4894 scope.go:117] "RemoveContainer" containerID="216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6" Dec 09 15:51:28 crc kubenswrapper[4894]: E1209 15:51:28.166329 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6\": container with ID starting with 216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6 not found: ID does not exist" containerID="216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.166355 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6"} err="failed to get container status \"216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6\": rpc error: code = NotFound desc = could not find container \"216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6\": container with ID starting with 216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6 not found: ID does not exist" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.166370 4894 scope.go:117] "RemoveContainer" containerID="99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3" Dec 09 15:51:28 crc kubenswrapper[4894]: E1209 15:51:28.166565 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3\": container with ID starting with 99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3 not found: ID does not exist" containerID="99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.166591 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3"} err="failed to get container status \"99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3\": rpc error: code = NotFound desc = could not find container \"99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3\": container with ID starting with 99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3 not found: ID does not exist" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.166606 4894 scope.go:117] "RemoveContainer" containerID="7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395" Dec 09 15:51:28 crc kubenswrapper[4894]: E1209 15:51:28.166831 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395\": container with ID starting with 7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395 not found: ID does not exist" containerID="7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.166857 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395"} err="failed to get container status \"7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395\": rpc error: code = NotFound desc = could not find container \"7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395\": container with ID starting with 7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395 not found: ID does not exist" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.166876 4894 scope.go:117] "RemoveContainer" containerID="e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.167057 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429"} err="failed to get container status \"e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429\": rpc error: code = NotFound desc = could not find container \"e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429\": container with ID starting with e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429 not found: ID does not exist" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.167079 4894 scope.go:117] "RemoveContainer" containerID="216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.167302 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6"} err="failed to get container status \"216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6\": rpc error: code = NotFound desc = could not find container \"216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6\": container with ID starting with 216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6 not found: ID does not exist" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.167323 4894 scope.go:117] "RemoveContainer" containerID="99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.167483 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3"} err="failed to get container status \"99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3\": rpc error: code = NotFound desc = could not find container \"99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3\": container with ID starting with 99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3 not found: ID does not exist" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.167503 4894 scope.go:117] "RemoveContainer" containerID="7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.167782 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395"} err="failed to get container status \"7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395\": rpc error: code = NotFound desc = could not find container \"7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395\": container with ID starting with 7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395 not found: ID does not exist" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.167806 4894 scope.go:117] "RemoveContainer" containerID="e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.168178 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429"} err="failed to get container status \"e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429\": rpc error: code = NotFound desc = could not find container \"e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429\": container with ID starting with e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429 not found: ID does not exist" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.168232 4894 scope.go:117] "RemoveContainer" containerID="216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.168503 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6"} err="failed to get container status \"216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6\": rpc error: code = NotFound desc = could not find container \"216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6\": container with ID starting with 216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6 not found: ID does not exist" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.168529 4894 scope.go:117] "RemoveContainer" containerID="99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.168791 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3"} err="failed to get container status \"99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3\": rpc error: code = NotFound desc = could not find container \"99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3\": container with ID starting with 99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3 not found: ID does not exist" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.168819 4894 scope.go:117] "RemoveContainer" containerID="7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.169799 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395"} err="failed to get container status \"7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395\": rpc error: code = NotFound desc = could not find container \"7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395\": container with ID starting with 7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395 not found: ID does not exist" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.169819 4894 scope.go:117] "RemoveContainer" containerID="e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.170740 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429"} err="failed to get container status \"e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429\": rpc error: code = NotFound desc = could not find container \"e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429\": container with ID starting with e9ce2379c1afa1f28acd3c60d1c306544db63c221802148ca3acf62edc1b4429 not found: ID does not exist" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.170778 4894 scope.go:117] "RemoveContainer" containerID="216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.171351 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6"} err="failed to get container status \"216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6\": rpc error: code = NotFound desc = could not find container \"216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6\": container with ID starting with 216f55a71776f9dd5770c3a0833090d319c85e3d64c307383d83b990d045e4e6 not found: ID does not exist" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.171446 4894 scope.go:117] "RemoveContainer" containerID="99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.171822 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3"} err="failed to get container status \"99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3\": rpc error: code = NotFound desc = could not find container \"99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3\": container with ID starting with 99663d61e2db735c753c3c96b3e2f9d2cd37685fee6d42bf5eb87685271b04d3 not found: ID does not exist" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.171846 4894 scope.go:117] "RemoveContainer" containerID="7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.172120 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395"} err="failed to get container status \"7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395\": rpc error: code = NotFound desc = could not find container \"7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395\": container with ID starting with 7543515fc769c8b1bc597f453a460c1ebb206ff508be3316e1ab4682817da395 not found: ID does not exist" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.234524 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.234580 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.234724 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-config-data\") pod \"ceilometer-0\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.234746 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fc523c0-a7fc-4689-a3ed-e43288589249-run-httpd\") pod \"ceilometer-0\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.234780 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-275rd\" (UniqueName: \"kubernetes.io/projected/7fc523c0-a7fc-4689-a3ed-e43288589249-kube-api-access-275rd\") pod \"ceilometer-0\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.234839 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-scripts\") pod \"ceilometer-0\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.234870 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fc523c0-a7fc-4689-a3ed-e43288589249-log-httpd\") pod \"ceilometer-0\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.336815 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fc523c0-a7fc-4689-a3ed-e43288589249-log-httpd\") pod \"ceilometer-0\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.337188 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.337227 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.337280 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-config-data\") pod \"ceilometer-0\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.337302 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fc523c0-a7fc-4689-a3ed-e43288589249-run-httpd\") pod \"ceilometer-0\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.337341 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-275rd\" (UniqueName: \"kubernetes.io/projected/7fc523c0-a7fc-4689-a3ed-e43288589249-kube-api-access-275rd\") pod \"ceilometer-0\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.337419 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-scripts\") pod \"ceilometer-0\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.337942 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fc523c0-a7fc-4689-a3ed-e43288589249-log-httpd\") pod \"ceilometer-0\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.338224 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fc523c0-a7fc-4689-a3ed-e43288589249-run-httpd\") pod \"ceilometer-0\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.342413 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.342413 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.343763 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-scripts\") pod \"ceilometer-0\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.345036 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-config-data\") pod \"ceilometer-0\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.358911 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-275rd\" (UniqueName: \"kubernetes.io/projected/7fc523c0-a7fc-4689-a3ed-e43288589249-kube-api-access-275rd\") pod \"ceilometer-0\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.456092 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:51:28 crc kubenswrapper[4894]: I1209 15:51:28.967602 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:51:28 crc kubenswrapper[4894]: W1209 15:51:28.978006 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fc523c0_a7fc_4689_a3ed_e43288589249.slice/crio-864a53b4fb77e885f8b056a0f8962402af3af1c26c8f441bc88311d8776769f7 WatchSource:0}: Error finding container 864a53b4fb77e885f8b056a0f8962402af3af1c26c8f441bc88311d8776769f7: Status 404 returned error can't find the container with id 864a53b4fb77e885f8b056a0f8962402af3af1c26c8f441bc88311d8776769f7 Dec 09 15:51:29 crc kubenswrapper[4894]: I1209 15:51:29.044483 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fc523c0-a7fc-4689-a3ed-e43288589249","Type":"ContainerStarted","Data":"864a53b4fb77e885f8b056a0f8962402af3af1c26c8f441bc88311d8776769f7"} Dec 09 15:51:30 crc kubenswrapper[4894]: I1209 15:51:30.057382 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fc523c0-a7fc-4689-a3ed-e43288589249","Type":"ContainerStarted","Data":"67eccfb195452e20fd8a6e68c0898294288d59b21488e44bb1ec66b34cf286f6"} Dec 09 15:51:30 crc kubenswrapper[4894]: I1209 15:51:30.118502 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="992661f7-3ddd-40fb-9790-268cc5c0b83f" path="/var/lib/kubelet/pods/992661f7-3ddd-40fb-9790-268cc5c0b83f/volumes" Dec 09 15:51:31 crc kubenswrapper[4894]: I1209 15:51:31.082841 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fc523c0-a7fc-4689-a3ed-e43288589249","Type":"ContainerStarted","Data":"df527646f6cc5b4c20571dd4d6a938bc0029167739028aac8ab7f819a664e895"} Dec 09 15:51:32 crc kubenswrapper[4894]: I1209 15:51:32.093803 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fc523c0-a7fc-4689-a3ed-e43288589249","Type":"ContainerStarted","Data":"860d30ad535d6172272c81311f33b40c1de0b026651c3282663f5518b92bb13a"} Dec 09 15:51:33 crc kubenswrapper[4894]: I1209 15:51:33.105415 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fc523c0-a7fc-4689-a3ed-e43288589249","Type":"ContainerStarted","Data":"7460daec013ac6ac4932f395ede5e82849994c5daabe2e8383db32ee12fc51a9"} Dec 09 15:51:33 crc kubenswrapper[4894]: I1209 15:51:33.105869 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 15:51:33 crc kubenswrapper[4894]: I1209 15:51:33.132797 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.787582717 podStartE2EDuration="5.132778368s" podCreationTimestamp="2025-12-09 15:51:28 +0000 UTC" firstStartedPulling="2025-12-09 15:51:28.98102065 +0000 UTC m=+1183.300231319" lastFinishedPulling="2025-12-09 15:51:32.326216301 +0000 UTC m=+1186.645426970" observedRunningTime="2025-12-09 15:51:33.126355055 +0000 UTC m=+1187.445565734" watchObservedRunningTime="2025-12-09 15:51:33.132778368 +0000 UTC m=+1187.451989037" Dec 09 15:51:34 crc kubenswrapper[4894]: I1209 15:51:34.116099 4894 generic.go:334] "Generic (PLEG): container finished" podID="8ed12da4-c785-4d59-a3d7-c485b97013bc" containerID="eb0e84f294bc81396d21a68308cbb6daf5f04026d580ce87f6bd2297b9c4fb4e" exitCode=0 Dec 09 15:51:34 crc kubenswrapper[4894]: I1209 15:51:34.118326 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-b56ck" event={"ID":"8ed12da4-c785-4d59-a3d7-c485b97013bc","Type":"ContainerDied","Data":"eb0e84f294bc81396d21a68308cbb6daf5f04026d580ce87f6bd2297b9c4fb4e"} Dec 09 15:51:35 crc kubenswrapper[4894]: I1209 15:51:35.488420 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-b56ck" Dec 09 15:51:35 crc kubenswrapper[4894]: I1209 15:51:35.584009 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ed12da4-c785-4d59-a3d7-c485b97013bc-scripts\") pod \"8ed12da4-c785-4d59-a3d7-c485b97013bc\" (UID: \"8ed12da4-c785-4d59-a3d7-c485b97013bc\") " Dec 09 15:51:35 crc kubenswrapper[4894]: I1209 15:51:35.584128 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ed12da4-c785-4d59-a3d7-c485b97013bc-combined-ca-bundle\") pod \"8ed12da4-c785-4d59-a3d7-c485b97013bc\" (UID: \"8ed12da4-c785-4d59-a3d7-c485b97013bc\") " Dec 09 15:51:35 crc kubenswrapper[4894]: I1209 15:51:35.584201 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ed12da4-c785-4d59-a3d7-c485b97013bc-config-data\") pod \"8ed12da4-c785-4d59-a3d7-c485b97013bc\" (UID: \"8ed12da4-c785-4d59-a3d7-c485b97013bc\") " Dec 09 15:51:35 crc kubenswrapper[4894]: I1209 15:51:35.584314 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gzhv\" (UniqueName: \"kubernetes.io/projected/8ed12da4-c785-4d59-a3d7-c485b97013bc-kube-api-access-5gzhv\") pod \"8ed12da4-c785-4d59-a3d7-c485b97013bc\" (UID: \"8ed12da4-c785-4d59-a3d7-c485b97013bc\") " Dec 09 15:51:35 crc kubenswrapper[4894]: I1209 15:51:35.603979 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8ed12da4-c785-4d59-a3d7-c485b97013bc-kube-api-access-5gzhv" (OuterVolumeSpecName: "kube-api-access-5gzhv") pod "8ed12da4-c785-4d59-a3d7-c485b97013bc" (UID: "8ed12da4-c785-4d59-a3d7-c485b97013bc"). InnerVolumeSpecName "kube-api-access-5gzhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:51:35 crc kubenswrapper[4894]: I1209 15:51:35.605695 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ed12da4-c785-4d59-a3d7-c485b97013bc-scripts" (OuterVolumeSpecName: "scripts") pod "8ed12da4-c785-4d59-a3d7-c485b97013bc" (UID: "8ed12da4-c785-4d59-a3d7-c485b97013bc"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:51:35 crc kubenswrapper[4894]: I1209 15:51:35.613728 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ed12da4-c785-4d59-a3d7-c485b97013bc-config-data" (OuterVolumeSpecName: "config-data") pod "8ed12da4-c785-4d59-a3d7-c485b97013bc" (UID: "8ed12da4-c785-4d59-a3d7-c485b97013bc"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:51:35 crc kubenswrapper[4894]: I1209 15:51:35.616150 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8ed12da4-c785-4d59-a3d7-c485b97013bc-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8ed12da4-c785-4d59-a3d7-c485b97013bc" (UID: "8ed12da4-c785-4d59-a3d7-c485b97013bc"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:51:35 crc kubenswrapper[4894]: I1209 15:51:35.686568 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gzhv\" (UniqueName: \"kubernetes.io/projected/8ed12da4-c785-4d59-a3d7-c485b97013bc-kube-api-access-5gzhv\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:35 crc kubenswrapper[4894]: I1209 15:51:35.686795 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8ed12da4-c785-4d59-a3d7-c485b97013bc-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:35 crc kubenswrapper[4894]: I1209 15:51:35.686856 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8ed12da4-c785-4d59-a3d7-c485b97013bc-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:35 crc kubenswrapper[4894]: I1209 15:51:35.686944 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8ed12da4-c785-4d59-a3d7-c485b97013bc-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:36 crc kubenswrapper[4894]: I1209 15:51:36.142540 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-b56ck" event={"ID":"8ed12da4-c785-4d59-a3d7-c485b97013bc","Type":"ContainerDied","Data":"a8485224bf9df9c87c412470d36c1071122e00b04afd66a918c84c4927d70dc4"} Dec 09 15:51:36 crc kubenswrapper[4894]: I1209 15:51:36.142595 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8485224bf9df9c87c412470d36c1071122e00b04afd66a918c84c4927d70dc4" Dec 09 15:51:36 crc kubenswrapper[4894]: I1209 15:51:36.142617 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-b56ck" Dec 09 15:51:36 crc kubenswrapper[4894]: I1209 15:51:36.251001 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 09 15:51:36 crc kubenswrapper[4894]: E1209 15:51:36.251432 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8ed12da4-c785-4d59-a3d7-c485b97013bc" containerName="nova-cell0-conductor-db-sync" Dec 09 15:51:36 crc kubenswrapper[4894]: I1209 15:51:36.251446 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8ed12da4-c785-4d59-a3d7-c485b97013bc" containerName="nova-cell0-conductor-db-sync" Dec 09 15:51:36 crc kubenswrapper[4894]: I1209 15:51:36.251698 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8ed12da4-c785-4d59-a3d7-c485b97013bc" containerName="nova-cell0-conductor-db-sync" Dec 09 15:51:36 crc kubenswrapper[4894]: I1209 15:51:36.252350 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 09 15:51:36 crc kubenswrapper[4894]: I1209 15:51:36.291418 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Dec 09 15:51:36 crc kubenswrapper[4894]: I1209 15:51:36.296943 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-6fv64" Dec 09 15:51:36 crc kubenswrapper[4894]: I1209 15:51:36.301426 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 09 15:51:36 crc kubenswrapper[4894]: I1209 15:51:36.400376 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8a9b33e-85ed-4c31-abb8-21bbca492891-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e8a9b33e-85ed-4c31-abb8-21bbca492891\") " pod="openstack/nova-cell0-conductor-0" Dec 09 15:51:36 crc kubenswrapper[4894]: I1209 15:51:36.400528 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8a9b33e-85ed-4c31-abb8-21bbca492891-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e8a9b33e-85ed-4c31-abb8-21bbca492891\") " pod="openstack/nova-cell0-conductor-0" Dec 09 15:51:36 crc kubenswrapper[4894]: I1209 15:51:36.400816 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmsks\" (UniqueName: \"kubernetes.io/projected/e8a9b33e-85ed-4c31-abb8-21bbca492891-kube-api-access-wmsks\") pod \"nova-cell0-conductor-0\" (UID: \"e8a9b33e-85ed-4c31-abb8-21bbca492891\") " pod="openstack/nova-cell0-conductor-0" Dec 09 15:51:36 crc kubenswrapper[4894]: I1209 15:51:36.502814 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmsks\" (UniqueName: \"kubernetes.io/projected/e8a9b33e-85ed-4c31-abb8-21bbca492891-kube-api-access-wmsks\") pod \"nova-cell0-conductor-0\" (UID: \"e8a9b33e-85ed-4c31-abb8-21bbca492891\") " pod="openstack/nova-cell0-conductor-0" Dec 09 15:51:36 crc kubenswrapper[4894]: I1209 15:51:36.502988 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8a9b33e-85ed-4c31-abb8-21bbca492891-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e8a9b33e-85ed-4c31-abb8-21bbca492891\") " pod="openstack/nova-cell0-conductor-0" Dec 09 15:51:36 crc kubenswrapper[4894]: I1209 15:51:36.503046 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8a9b33e-85ed-4c31-abb8-21bbca492891-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e8a9b33e-85ed-4c31-abb8-21bbca492891\") " pod="openstack/nova-cell0-conductor-0" Dec 09 15:51:36 crc kubenswrapper[4894]: I1209 15:51:36.515499 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8a9b33e-85ed-4c31-abb8-21bbca492891-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"e8a9b33e-85ed-4c31-abb8-21bbca492891\") " pod="openstack/nova-cell0-conductor-0" Dec 09 15:51:36 crc kubenswrapper[4894]: I1209 15:51:36.528528 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8a9b33e-85ed-4c31-abb8-21bbca492891-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"e8a9b33e-85ed-4c31-abb8-21bbca492891\") " pod="openstack/nova-cell0-conductor-0" Dec 09 15:51:36 crc kubenswrapper[4894]: I1209 15:51:36.537048 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmsks\" (UniqueName: \"kubernetes.io/projected/e8a9b33e-85ed-4c31-abb8-21bbca492891-kube-api-access-wmsks\") pod \"nova-cell0-conductor-0\" (UID: \"e8a9b33e-85ed-4c31-abb8-21bbca492891\") " pod="openstack/nova-cell0-conductor-0" Dec 09 15:51:36 crc kubenswrapper[4894]: I1209 15:51:36.609241 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Dec 09 15:51:37 crc kubenswrapper[4894]: I1209 15:51:37.074962 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Dec 09 15:51:37 crc kubenswrapper[4894]: W1209 15:51:37.084849 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8a9b33e_85ed_4c31_abb8_21bbca492891.slice/crio-6ba77bb368adae4d38f2395dcffd621474e7d95d9eee1559e11e2194878b21c1 WatchSource:0}: Error finding container 6ba77bb368adae4d38f2395dcffd621474e7d95d9eee1559e11e2194878b21c1: Status 404 returned error can't find the container with id 6ba77bb368adae4d38f2395dcffd621474e7d95d9eee1559e11e2194878b21c1 Dec 09 15:51:37 crc kubenswrapper[4894]: I1209 15:51:37.156707 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e8a9b33e-85ed-4c31-abb8-21bbca492891","Type":"ContainerStarted","Data":"6ba77bb368adae4d38f2395dcffd621474e7d95d9eee1559e11e2194878b21c1"} Dec 09 15:51:38 crc kubenswrapper[4894]: I1209 15:51:38.173187 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"e8a9b33e-85ed-4c31-abb8-21bbca492891","Type":"ContainerStarted","Data":"7487545fe94b31c3c72cfae222c4e74a3c85749854873e72072fbfbacb62120f"} Dec 09 15:51:38 crc kubenswrapper[4894]: I1209 15:51:38.174561 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Dec 09 15:51:38 crc kubenswrapper[4894]: I1209 15:51:38.197900 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=2.197874904 podStartE2EDuration="2.197874904s" podCreationTimestamp="2025-12-09 15:51:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:51:38.190262132 +0000 UTC m=+1192.509472801" watchObservedRunningTime="2025-12-09 15:51:38.197874904 +0000 UTC m=+1192.517085573" Dec 09 15:51:46 crc kubenswrapper[4894]: I1209 15:51:46.637563 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.046726 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-w26bv"] Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.048283 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-w26bv" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.050896 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.051512 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.058629 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-w26bv"] Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.201499 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-scripts\") pod \"nova-cell0-cell-mapping-w26bv\" (UID: \"e2f0b872-8eba-4309-94c8-4dfe893cfc3c\") " pod="openstack/nova-cell0-cell-mapping-w26bv" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.201609 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-w26bv\" (UID: \"e2f0b872-8eba-4309-94c8-4dfe893cfc3c\") " pod="openstack/nova-cell0-cell-mapping-w26bv" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.201699 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-config-data\") pod \"nova-cell0-cell-mapping-w26bv\" (UID: \"e2f0b872-8eba-4309-94c8-4dfe893cfc3c\") " pod="openstack/nova-cell0-cell-mapping-w26bv" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.201734 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jx6dg\" (UniqueName: \"kubernetes.io/projected/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-kube-api-access-jx6dg\") pod \"nova-cell0-cell-mapping-w26bv\" (UID: \"e2f0b872-8eba-4309-94c8-4dfe893cfc3c\") " pod="openstack/nova-cell0-cell-mapping-w26bv" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.215313 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.216765 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.222538 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.226764 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.297212 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.298427 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.303192 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wjjf\" (UniqueName: \"kubernetes.io/projected/ee76fef6-16b9-49e7-adb6-2145cc0e5e03-kube-api-access-6wjjf\") pod \"nova-cell1-novncproxy-0\" (UID: \"ee76fef6-16b9-49e7-adb6-2145cc0e5e03\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.303285 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-config-data\") pod \"nova-cell0-cell-mapping-w26bv\" (UID: \"e2f0b872-8eba-4309-94c8-4dfe893cfc3c\") " pod="openstack/nova-cell0-cell-mapping-w26bv" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.303333 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jx6dg\" (UniqueName: \"kubernetes.io/projected/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-kube-api-access-jx6dg\") pod \"nova-cell0-cell-mapping-w26bv\" (UID: \"e2f0b872-8eba-4309-94c8-4dfe893cfc3c\") " pod="openstack/nova-cell0-cell-mapping-w26bv" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.303397 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-scripts\") pod \"nova-cell0-cell-mapping-w26bv\" (UID: \"e2f0b872-8eba-4309-94c8-4dfe893cfc3c\") " pod="openstack/nova-cell0-cell-mapping-w26bv" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.303484 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee76fef6-16b9-49e7-adb6-2145cc0e5e03-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ee76fef6-16b9-49e7-adb6-2145cc0e5e03\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.303519 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee76fef6-16b9-49e7-adb6-2145cc0e5e03-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ee76fef6-16b9-49e7-adb6-2145cc0e5e03\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.303574 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-w26bv\" (UID: \"e2f0b872-8eba-4309-94c8-4dfe893cfc3c\") " pod="openstack/nova-cell0-cell-mapping-w26bv" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.305589 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.308045 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.315602 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-config-data\") pod \"nova-cell0-cell-mapping-w26bv\" (UID: \"e2f0b872-8eba-4309-94c8-4dfe893cfc3c\") " pod="openstack/nova-cell0-cell-mapping-w26bv" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.316052 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-scripts\") pod \"nova-cell0-cell-mapping-w26bv\" (UID: \"e2f0b872-8eba-4309-94c8-4dfe893cfc3c\") " pod="openstack/nova-cell0-cell-mapping-w26bv" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.319931 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-w26bv\" (UID: \"e2f0b872-8eba-4309-94c8-4dfe893cfc3c\") " pod="openstack/nova-cell0-cell-mapping-w26bv" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.385341 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jx6dg\" (UniqueName: \"kubernetes.io/projected/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-kube-api-access-jx6dg\") pod \"nova-cell0-cell-mapping-w26bv\" (UID: \"e2f0b872-8eba-4309-94c8-4dfe893cfc3c\") " pod="openstack/nova-cell0-cell-mapping-w26bv" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.404973 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cefa8f8-0c44-4827-af46-671cea04dd97-config-data\") pod \"nova-scheduler-0\" (UID: \"8cefa8f8-0c44-4827-af46-671cea04dd97\") " pod="openstack/nova-scheduler-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.405017 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gv979\" (UniqueName: \"kubernetes.io/projected/8cefa8f8-0c44-4827-af46-671cea04dd97-kube-api-access-gv979\") pod \"nova-scheduler-0\" (UID: \"8cefa8f8-0c44-4827-af46-671cea04dd97\") " pod="openstack/nova-scheduler-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.405059 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee76fef6-16b9-49e7-adb6-2145cc0e5e03-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ee76fef6-16b9-49e7-adb6-2145cc0e5e03\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.405105 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee76fef6-16b9-49e7-adb6-2145cc0e5e03-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ee76fef6-16b9-49e7-adb6-2145cc0e5e03\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.405174 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wjjf\" (UniqueName: \"kubernetes.io/projected/ee76fef6-16b9-49e7-adb6-2145cc0e5e03-kube-api-access-6wjjf\") pod \"nova-cell1-novncproxy-0\" (UID: \"ee76fef6-16b9-49e7-adb6-2145cc0e5e03\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.405221 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cefa8f8-0c44-4827-af46-671cea04dd97-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8cefa8f8-0c44-4827-af46-671cea04dd97\") " pod="openstack/nova-scheduler-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.409044 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee76fef6-16b9-49e7-adb6-2145cc0e5e03-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"ee76fef6-16b9-49e7-adb6-2145cc0e5e03\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.410667 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee76fef6-16b9-49e7-adb6-2145cc0e5e03-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"ee76fef6-16b9-49e7-adb6-2145cc0e5e03\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.417526 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.419395 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.426137 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.452010 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.464968 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wjjf\" (UniqueName: \"kubernetes.io/projected/ee76fef6-16b9-49e7-adb6-2145cc0e5e03-kube-api-access-6wjjf\") pod \"nova-cell1-novncproxy-0\" (UID: \"ee76fef6-16b9-49e7-adb6-2145cc0e5e03\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.508662 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v89v4\" (UniqueName: \"kubernetes.io/projected/804394f9-a48a-4e95-acc3-116261551091-kube-api-access-v89v4\") pod \"nova-metadata-0\" (UID: \"804394f9-a48a-4e95-acc3-116261551091\") " pod="openstack/nova-metadata-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.508715 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/804394f9-a48a-4e95-acc3-116261551091-logs\") pod \"nova-metadata-0\" (UID: \"804394f9-a48a-4e95-acc3-116261551091\") " pod="openstack/nova-metadata-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.508768 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cefa8f8-0c44-4827-af46-671cea04dd97-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8cefa8f8-0c44-4827-af46-671cea04dd97\") " pod="openstack/nova-scheduler-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.508803 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/804394f9-a48a-4e95-acc3-116261551091-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"804394f9-a48a-4e95-acc3-116261551091\") " pod="openstack/nova-metadata-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.508854 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cefa8f8-0c44-4827-af46-671cea04dd97-config-data\") pod \"nova-scheduler-0\" (UID: \"8cefa8f8-0c44-4827-af46-671cea04dd97\") " pod="openstack/nova-scheduler-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.508871 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gv979\" (UniqueName: \"kubernetes.io/projected/8cefa8f8-0c44-4827-af46-671cea04dd97-kube-api-access-gv979\") pod \"nova-scheduler-0\" (UID: \"8cefa8f8-0c44-4827-af46-671cea04dd97\") " pod="openstack/nova-scheduler-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.508894 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/804394f9-a48a-4e95-acc3-116261551091-config-data\") pod \"nova-metadata-0\" (UID: \"804394f9-a48a-4e95-acc3-116261551091\") " pod="openstack/nova-metadata-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.513222 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cefa8f8-0c44-4827-af46-671cea04dd97-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"8cefa8f8-0c44-4827-af46-671cea04dd97\") " pod="openstack/nova-scheduler-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.519592 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.520835 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cefa8f8-0c44-4827-af46-671cea04dd97-config-data\") pod \"nova-scheduler-0\" (UID: \"8cefa8f8-0c44-4827-af46-671cea04dd97\") " pod="openstack/nova-scheduler-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.521708 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.534601 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.538038 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.566455 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gv979\" (UniqueName: \"kubernetes.io/projected/8cefa8f8-0c44-4827-af46-671cea04dd97-kube-api-access-gv979\") pod \"nova-scheduler-0\" (UID: \"8cefa8f8-0c44-4827-af46-671cea04dd97\") " pod="openstack/nova-scheduler-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.577719 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8d97cbc7-jdzrx"] Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.587315 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.592373 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.607741 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8d97cbc7-jdzrx"] Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.611393 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9418641c-296a-4d30-a02a-52c150f99cd2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9418641c-296a-4d30-a02a-52c150f99cd2\") " pod="openstack/nova-api-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.611485 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/804394f9-a48a-4e95-acc3-116261551091-config-data\") pod \"nova-metadata-0\" (UID: \"804394f9-a48a-4e95-acc3-116261551091\") " pod="openstack/nova-metadata-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.611529 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9418641c-296a-4d30-a02a-52c150f99cd2-logs\") pod \"nova-api-0\" (UID: \"9418641c-296a-4d30-a02a-52c150f99cd2\") " pod="openstack/nova-api-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.611554 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9418641c-296a-4d30-a02a-52c150f99cd2-config-data\") pod \"nova-api-0\" (UID: \"9418641c-296a-4d30-a02a-52c150f99cd2\") " pod="openstack/nova-api-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.611839 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v89v4\" (UniqueName: \"kubernetes.io/projected/804394f9-a48a-4e95-acc3-116261551091-kube-api-access-v89v4\") pod \"nova-metadata-0\" (UID: \"804394f9-a48a-4e95-acc3-116261551091\") " pod="openstack/nova-metadata-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.611874 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/804394f9-a48a-4e95-acc3-116261551091-logs\") pod \"nova-metadata-0\" (UID: \"804394f9-a48a-4e95-acc3-116261551091\") " pod="openstack/nova-metadata-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.611920 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swcwt\" (UniqueName: \"kubernetes.io/projected/9418641c-296a-4d30-a02a-52c150f99cd2-kube-api-access-swcwt\") pod \"nova-api-0\" (UID: \"9418641c-296a-4d30-a02a-52c150f99cd2\") " pod="openstack/nova-api-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.612058 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/804394f9-a48a-4e95-acc3-116261551091-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"804394f9-a48a-4e95-acc3-116261551091\") " pod="openstack/nova-metadata-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.613467 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/804394f9-a48a-4e95-acc3-116261551091-logs\") pod \"nova-metadata-0\" (UID: \"804394f9-a48a-4e95-acc3-116261551091\") " pod="openstack/nova-metadata-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.615244 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/804394f9-a48a-4e95-acc3-116261551091-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"804394f9-a48a-4e95-acc3-116261551091\") " pod="openstack/nova-metadata-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.618731 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/804394f9-a48a-4e95-acc3-116261551091-config-data\") pod \"nova-metadata-0\" (UID: \"804394f9-a48a-4e95-acc3-116261551091\") " pod="openstack/nova-metadata-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.635753 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v89v4\" (UniqueName: \"kubernetes.io/projected/804394f9-a48a-4e95-acc3-116261551091-kube-api-access-v89v4\") pod \"nova-metadata-0\" (UID: \"804394f9-a48a-4e95-acc3-116261551091\") " pod="openstack/nova-metadata-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.668012 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-w26bv" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.730253 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-ovsdbserver-sb\") pod \"dnsmasq-dns-8d97cbc7-jdzrx\" (UID: \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\") " pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.730322 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swcwt\" (UniqueName: \"kubernetes.io/projected/9418641c-296a-4d30-a02a-52c150f99cd2-kube-api-access-swcwt\") pod \"nova-api-0\" (UID: \"9418641c-296a-4d30-a02a-52c150f99cd2\") " pod="openstack/nova-api-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.730417 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-config\") pod \"dnsmasq-dns-8d97cbc7-jdzrx\" (UID: \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\") " pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.730442 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9418641c-296a-4d30-a02a-52c150f99cd2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9418641c-296a-4d30-a02a-52c150f99cd2\") " pod="openstack/nova-api-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.730482 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-ovsdbserver-nb\") pod \"dnsmasq-dns-8d97cbc7-jdzrx\" (UID: \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\") " pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.730528 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-dns-svc\") pod \"dnsmasq-dns-8d97cbc7-jdzrx\" (UID: \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\") " pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.730561 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9418641c-296a-4d30-a02a-52c150f99cd2-logs\") pod \"nova-api-0\" (UID: \"9418641c-296a-4d30-a02a-52c150f99cd2\") " pod="openstack/nova-api-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.730584 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9418641c-296a-4d30-a02a-52c150f99cd2-config-data\") pod \"nova-api-0\" (UID: \"9418641c-296a-4d30-a02a-52c150f99cd2\") " pod="openstack/nova-api-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.730612 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl4wb\" (UniqueName: \"kubernetes.io/projected/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-kube-api-access-gl4wb\") pod \"dnsmasq-dns-8d97cbc7-jdzrx\" (UID: \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\") " pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.732973 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9418641c-296a-4d30-a02a-52c150f99cd2-logs\") pod \"nova-api-0\" (UID: \"9418641c-296a-4d30-a02a-52c150f99cd2\") " pod="openstack/nova-api-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.734973 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9418641c-296a-4d30-a02a-52c150f99cd2-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9418641c-296a-4d30-a02a-52c150f99cd2\") " pod="openstack/nova-api-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.736683 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9418641c-296a-4d30-a02a-52c150f99cd2-config-data\") pod \"nova-api-0\" (UID: \"9418641c-296a-4d30-a02a-52c150f99cd2\") " pod="openstack/nova-api-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.755485 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swcwt\" (UniqueName: \"kubernetes.io/projected/9418641c-296a-4d30-a02a-52c150f99cd2-kube-api-access-swcwt\") pod \"nova-api-0\" (UID: \"9418641c-296a-4d30-a02a-52c150f99cd2\") " pod="openstack/nova-api-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.764245 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.836563 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-ovsdbserver-sb\") pod \"dnsmasq-dns-8d97cbc7-jdzrx\" (UID: \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\") " pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.836727 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-config\") pod \"dnsmasq-dns-8d97cbc7-jdzrx\" (UID: \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\") " pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.836777 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-ovsdbserver-nb\") pod \"dnsmasq-dns-8d97cbc7-jdzrx\" (UID: \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\") " pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.836817 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-dns-svc\") pod \"dnsmasq-dns-8d97cbc7-jdzrx\" (UID: \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\") " pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.836858 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl4wb\" (UniqueName: \"kubernetes.io/projected/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-kube-api-access-gl4wb\") pod \"dnsmasq-dns-8d97cbc7-jdzrx\" (UID: \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\") " pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.839933 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-dns-svc\") pod \"dnsmasq-dns-8d97cbc7-jdzrx\" (UID: \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\") " pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.839840 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-ovsdbserver-sb\") pod \"dnsmasq-dns-8d97cbc7-jdzrx\" (UID: \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\") " pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.840704 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-config\") pod \"dnsmasq-dns-8d97cbc7-jdzrx\" (UID: \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\") " pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.841298 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-ovsdbserver-nb\") pod \"dnsmasq-dns-8d97cbc7-jdzrx\" (UID: \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\") " pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.855273 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.863119 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl4wb\" (UniqueName: \"kubernetes.io/projected/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-kube-api-access-gl4wb\") pod \"dnsmasq-dns-8d97cbc7-jdzrx\" (UID: \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\") " pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.889357 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 15:51:47 crc kubenswrapper[4894]: I1209 15:51:47.922657 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.191329 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.215095 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-scszh"] Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.216317 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-scszh" Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.218499 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.219916 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.251595 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-scszh"] Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.284834 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ee76fef6-16b9-49e7-adb6-2145cc0e5e03","Type":"ContainerStarted","Data":"17f38d262e9e0e9200c407f0f8c07a7276297df704f7c6565e8574e5f1d33a6f"} Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.290690 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-w26bv"] Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.352807 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jk65l\" (UniqueName: \"kubernetes.io/projected/0f261fad-179c-4f25-bc42-44691e48be50-kube-api-access-jk65l\") pod \"nova-cell1-conductor-db-sync-scszh\" (UID: \"0f261fad-179c-4f25-bc42-44691e48be50\") " pod="openstack/nova-cell1-conductor-db-sync-scszh" Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.353276 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f261fad-179c-4f25-bc42-44691e48be50-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-scszh\" (UID: \"0f261fad-179c-4f25-bc42-44691e48be50\") " pod="openstack/nova-cell1-conductor-db-sync-scszh" Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.353418 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f261fad-179c-4f25-bc42-44691e48be50-scripts\") pod \"nova-cell1-conductor-db-sync-scszh\" (UID: \"0f261fad-179c-4f25-bc42-44691e48be50\") " pod="openstack/nova-cell1-conductor-db-sync-scszh" Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.353495 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f261fad-179c-4f25-bc42-44691e48be50-config-data\") pod \"nova-cell1-conductor-db-sync-scszh\" (UID: \"0f261fad-179c-4f25-bc42-44691e48be50\") " pod="openstack/nova-cell1-conductor-db-sync-scszh" Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.372830 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.455807 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jk65l\" (UniqueName: \"kubernetes.io/projected/0f261fad-179c-4f25-bc42-44691e48be50-kube-api-access-jk65l\") pod \"nova-cell1-conductor-db-sync-scszh\" (UID: \"0f261fad-179c-4f25-bc42-44691e48be50\") " pod="openstack/nova-cell1-conductor-db-sync-scszh" Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.456022 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f261fad-179c-4f25-bc42-44691e48be50-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-scszh\" (UID: \"0f261fad-179c-4f25-bc42-44691e48be50\") " pod="openstack/nova-cell1-conductor-db-sync-scszh" Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.456069 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f261fad-179c-4f25-bc42-44691e48be50-scripts\") pod \"nova-cell1-conductor-db-sync-scszh\" (UID: \"0f261fad-179c-4f25-bc42-44691e48be50\") " pod="openstack/nova-cell1-conductor-db-sync-scszh" Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.456098 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f261fad-179c-4f25-bc42-44691e48be50-config-data\") pod \"nova-cell1-conductor-db-sync-scszh\" (UID: \"0f261fad-179c-4f25-bc42-44691e48be50\") " pod="openstack/nova-cell1-conductor-db-sync-scszh" Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.460737 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f261fad-179c-4f25-bc42-44691e48be50-scripts\") pod \"nova-cell1-conductor-db-sync-scszh\" (UID: \"0f261fad-179c-4f25-bc42-44691e48be50\") " pod="openstack/nova-cell1-conductor-db-sync-scszh" Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.460766 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f261fad-179c-4f25-bc42-44691e48be50-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-scszh\" (UID: \"0f261fad-179c-4f25-bc42-44691e48be50\") " pod="openstack/nova-cell1-conductor-db-sync-scszh" Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.465945 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f261fad-179c-4f25-bc42-44691e48be50-config-data\") pod \"nova-cell1-conductor-db-sync-scszh\" (UID: \"0f261fad-179c-4f25-bc42-44691e48be50\") " pod="openstack/nova-cell1-conductor-db-sync-scszh" Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.473054 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jk65l\" (UniqueName: \"kubernetes.io/projected/0f261fad-179c-4f25-bc42-44691e48be50-kube-api-access-jk65l\") pod \"nova-cell1-conductor-db-sync-scszh\" (UID: \"0f261fad-179c-4f25-bc42-44691e48be50\") " pod="openstack/nova-cell1-conductor-db-sync-scszh" Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.560247 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-scszh" Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.597889 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.650164 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 15:51:48 crc kubenswrapper[4894]: I1209 15:51:48.710497 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8d97cbc7-jdzrx"] Dec 09 15:51:49 crc kubenswrapper[4894]: I1209 15:51:49.237042 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-scszh"] Dec 09 15:51:49 crc kubenswrapper[4894]: W1209 15:51:49.249466 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0f261fad_179c_4f25_bc42_44691e48be50.slice/crio-59dd9db98efca6ebed794359db462ea823df28e1411babd70f8d682850b5da2f WatchSource:0}: Error finding container 59dd9db98efca6ebed794359db462ea823df28e1411babd70f8d682850b5da2f: Status 404 returned error can't find the container with id 59dd9db98efca6ebed794359db462ea823df28e1411babd70f8d682850b5da2f Dec 09 15:51:49 crc kubenswrapper[4894]: I1209 15:51:49.304216 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"804394f9-a48a-4e95-acc3-116261551091","Type":"ContainerStarted","Data":"15bf1587a4d0691b9528084b312b389f0f6f0af31fc1e278d1f7865d82091dc7"} Dec 09 15:51:49 crc kubenswrapper[4894]: I1209 15:51:49.308873 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-scszh" event={"ID":"0f261fad-179c-4f25-bc42-44691e48be50","Type":"ContainerStarted","Data":"59dd9db98efca6ebed794359db462ea823df28e1411babd70f8d682850b5da2f"} Dec 09 15:51:49 crc kubenswrapper[4894]: I1209 15:51:49.310945 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9418641c-296a-4d30-a02a-52c150f99cd2","Type":"ContainerStarted","Data":"b05d10557cd9a98666a0d567674b620320fbc19d2bcef8c03149816bd99aaa14"} Dec 09 15:51:49 crc kubenswrapper[4894]: I1209 15:51:49.317862 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-w26bv" event={"ID":"e2f0b872-8eba-4309-94c8-4dfe893cfc3c","Type":"ContainerStarted","Data":"389f6a7420333dcd236202ca16cd594b4e9bb3dd3e8434999e9784a6d984c7c9"} Dec 09 15:51:49 crc kubenswrapper[4894]: I1209 15:51:49.317944 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-w26bv" event={"ID":"e2f0b872-8eba-4309-94c8-4dfe893cfc3c","Type":"ContainerStarted","Data":"0640508dec8da6be94969fbbef9bffa9766bd27992ad1a305924706a40a987ff"} Dec 09 15:51:49 crc kubenswrapper[4894]: I1209 15:51:49.333354 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8cefa8f8-0c44-4827-af46-671cea04dd97","Type":"ContainerStarted","Data":"e45489f2964965311cd4e86cc33d4f71da390cc4c562f46b582e2cd534322a83"} Dec 09 15:51:49 crc kubenswrapper[4894]: I1209 15:51:49.335971 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-w26bv" podStartSLOduration=2.335951559 podStartE2EDuration="2.335951559s" podCreationTimestamp="2025-12-09 15:51:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:51:49.333916478 +0000 UTC m=+1203.653127147" watchObservedRunningTime="2025-12-09 15:51:49.335951559 +0000 UTC m=+1203.655162228" Dec 09 15:51:49 crc kubenswrapper[4894]: I1209 15:51:49.359326 4894 generic.go:334] "Generic (PLEG): container finished" podID="f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5" containerID="09d4925167d5ad7e670ed67ebb75fd150a844f525762c0ae9cbb92bf77b00b2f" exitCode=0 Dec 09 15:51:49 crc kubenswrapper[4894]: I1209 15:51:49.359613 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" event={"ID":"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5","Type":"ContainerDied","Data":"09d4925167d5ad7e670ed67ebb75fd150a844f525762c0ae9cbb92bf77b00b2f"} Dec 09 15:51:49 crc kubenswrapper[4894]: I1209 15:51:49.359658 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" event={"ID":"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5","Type":"ContainerStarted","Data":"c4a267a612dcb97055ff6cf0c4470d0463a41d78a809b8cb0956117b58c226c0"} Dec 09 15:51:50 crc kubenswrapper[4894]: I1209 15:51:50.371944 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-scszh" event={"ID":"0f261fad-179c-4f25-bc42-44691e48be50","Type":"ContainerStarted","Data":"29e9cc197d569126cf8be32f02028738bb59d1c4167b6af1d9d999aecd711a2e"} Dec 09 15:51:50 crc kubenswrapper[4894]: I1209 15:51:50.387586 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-scszh" podStartSLOduration=2.387566047 podStartE2EDuration="2.387566047s" podCreationTimestamp="2025-12-09 15:51:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:51:50.386618243 +0000 UTC m=+1204.705828912" watchObservedRunningTime="2025-12-09 15:51:50.387566047 +0000 UTC m=+1204.706776736" Dec 09 15:51:51 crc kubenswrapper[4894]: I1209 15:51:51.782845 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 15:51:51 crc kubenswrapper[4894]: I1209 15:51:51.791607 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 15:51:54 crc kubenswrapper[4894]: I1209 15:51:54.442701 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9418641c-296a-4d30-a02a-52c150f99cd2","Type":"ContainerStarted","Data":"ec90db2fbe4702580bec2f02502f736637c41093f6c7f277c82047e5b880698f"} Dec 09 15:51:54 crc kubenswrapper[4894]: I1209 15:51:54.454628 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8cefa8f8-0c44-4827-af46-671cea04dd97","Type":"ContainerStarted","Data":"f2556776659526f969cb1e90d6261ee3f6e0696cae9699946f744b0fac60b09b"} Dec 09 15:51:54 crc kubenswrapper[4894]: I1209 15:51:54.458335 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" event={"ID":"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5","Type":"ContainerStarted","Data":"49883e5ea70af46c31fe682fee954dba8d75f41207f6cc81ea8f694214b908c9"} Dec 09 15:51:54 crc kubenswrapper[4894]: I1209 15:51:54.458530 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" Dec 09 15:51:54 crc kubenswrapper[4894]: I1209 15:51:54.460708 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ee76fef6-16b9-49e7-adb6-2145cc0e5e03","Type":"ContainerStarted","Data":"2633f428a1d9bbb3c94cbe02ce3dc5ad819ef17b00d7689b8ead36a81c92ec74"} Dec 09 15:51:54 crc kubenswrapper[4894]: I1209 15:51:54.460905 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="ee76fef6-16b9-49e7-adb6-2145cc0e5e03" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://2633f428a1d9bbb3c94cbe02ce3dc5ad819ef17b00d7689b8ead36a81c92ec74" gracePeriod=30 Dec 09 15:51:54 crc kubenswrapper[4894]: I1209 15:51:54.467866 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"804394f9-a48a-4e95-acc3-116261551091","Type":"ContainerStarted","Data":"07ab85febb0174b07422933bf634288d5ebc79aee758a336123a62e63063c133"} Dec 09 15:51:54 crc kubenswrapper[4894]: I1209 15:51:54.480861 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=1.904076183 podStartE2EDuration="7.480835955s" podCreationTimestamp="2025-12-09 15:51:47 +0000 UTC" firstStartedPulling="2025-12-09 15:51:48.401490106 +0000 UTC m=+1202.720700775" lastFinishedPulling="2025-12-09 15:51:53.978249858 +0000 UTC m=+1208.297460547" observedRunningTime="2025-12-09 15:51:54.470463473 +0000 UTC m=+1208.789674142" watchObservedRunningTime="2025-12-09 15:51:54.480835955 +0000 UTC m=+1208.800046624" Dec 09 15:51:54 crc kubenswrapper[4894]: I1209 15:51:54.497611 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.702767369 podStartE2EDuration="7.497593309s" podCreationTimestamp="2025-12-09 15:51:47 +0000 UTC" firstStartedPulling="2025-12-09 15:51:48.179622812 +0000 UTC m=+1202.498833481" lastFinishedPulling="2025-12-09 15:51:53.974448752 +0000 UTC m=+1208.293659421" observedRunningTime="2025-12-09 15:51:54.494511281 +0000 UTC m=+1208.813721950" watchObservedRunningTime="2025-12-09 15:51:54.497593309 +0000 UTC m=+1208.816803978" Dec 09 15:51:55 crc kubenswrapper[4894]: I1209 15:51:55.478268 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9418641c-296a-4d30-a02a-52c150f99cd2","Type":"ContainerStarted","Data":"8687a54d3c71b9460c2c6bfe89afe97dd96d5b812daf0a197b0df3633e5b6c66"} Dec 09 15:51:55 crc kubenswrapper[4894]: I1209 15:51:55.479800 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"804394f9-a48a-4e95-acc3-116261551091","Type":"ContainerStarted","Data":"4c18fb89fb255d15b9bf0fba6f95c842299b7c92e911ad69bf975428c012a38a"} Dec 09 15:51:55 crc kubenswrapper[4894]: I1209 15:51:55.480257 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="804394f9-a48a-4e95-acc3-116261551091" containerName="nova-metadata-log" containerID="cri-o://07ab85febb0174b07422933bf634288d5ebc79aee758a336123a62e63063c133" gracePeriod=30 Dec 09 15:51:55 crc kubenswrapper[4894]: I1209 15:51:55.480424 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="804394f9-a48a-4e95-acc3-116261551091" containerName="nova-metadata-metadata" containerID="cri-o://4c18fb89fb255d15b9bf0fba6f95c842299b7c92e911ad69bf975428c012a38a" gracePeriod=30 Dec 09 15:51:55 crc kubenswrapper[4894]: I1209 15:51:55.499043 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" podStartSLOduration=8.499022777 podStartE2EDuration="8.499022777s" podCreationTimestamp="2025-12-09 15:51:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:51:54.512574088 +0000 UTC m=+1208.831784767" watchObservedRunningTime="2025-12-09 15:51:55.499022777 +0000 UTC m=+1209.818233446" Dec 09 15:51:55 crc kubenswrapper[4894]: I1209 15:51:55.499483 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.204901114 podStartE2EDuration="8.499477638s" podCreationTimestamp="2025-12-09 15:51:47 +0000 UTC" firstStartedPulling="2025-12-09 15:51:48.681820038 +0000 UTC m=+1203.001030707" lastFinishedPulling="2025-12-09 15:51:53.976396552 +0000 UTC m=+1208.295607231" observedRunningTime="2025-12-09 15:51:55.493796854 +0000 UTC m=+1209.813007523" watchObservedRunningTime="2025-12-09 15:51:55.499477638 +0000 UTC m=+1209.818688307" Dec 09 15:51:55 crc kubenswrapper[4894]: I1209 15:51:55.520372 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=3.161315083 podStartE2EDuration="8.520352036s" podCreationTimestamp="2025-12-09 15:51:47 +0000 UTC" firstStartedPulling="2025-12-09 15:51:48.615345067 +0000 UTC m=+1202.934555736" lastFinishedPulling="2025-12-09 15:51:53.97438202 +0000 UTC m=+1208.293592689" observedRunningTime="2025-12-09 15:51:55.51812024 +0000 UTC m=+1209.837330929" watchObservedRunningTime="2025-12-09 15:51:55.520352036 +0000 UTC m=+1209.839562705" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.080538 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.217040 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/804394f9-a48a-4e95-acc3-116261551091-logs\") pod \"804394f9-a48a-4e95-acc3-116261551091\" (UID: \"804394f9-a48a-4e95-acc3-116261551091\") " Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.217397 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/804394f9-a48a-4e95-acc3-116261551091-combined-ca-bundle\") pod \"804394f9-a48a-4e95-acc3-116261551091\" (UID: \"804394f9-a48a-4e95-acc3-116261551091\") " Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.217451 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/804394f9-a48a-4e95-acc3-116261551091-config-data\") pod \"804394f9-a48a-4e95-acc3-116261551091\" (UID: \"804394f9-a48a-4e95-acc3-116261551091\") " Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.217463 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/804394f9-a48a-4e95-acc3-116261551091-logs" (OuterVolumeSpecName: "logs") pod "804394f9-a48a-4e95-acc3-116261551091" (UID: "804394f9-a48a-4e95-acc3-116261551091"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.217506 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v89v4\" (UniqueName: \"kubernetes.io/projected/804394f9-a48a-4e95-acc3-116261551091-kube-api-access-v89v4\") pod \"804394f9-a48a-4e95-acc3-116261551091\" (UID: \"804394f9-a48a-4e95-acc3-116261551091\") " Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.217916 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/804394f9-a48a-4e95-acc3-116261551091-logs\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.228087 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/804394f9-a48a-4e95-acc3-116261551091-kube-api-access-v89v4" (OuterVolumeSpecName: "kube-api-access-v89v4") pod "804394f9-a48a-4e95-acc3-116261551091" (UID: "804394f9-a48a-4e95-acc3-116261551091"). InnerVolumeSpecName "kube-api-access-v89v4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.250181 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/804394f9-a48a-4e95-acc3-116261551091-config-data" (OuterVolumeSpecName: "config-data") pod "804394f9-a48a-4e95-acc3-116261551091" (UID: "804394f9-a48a-4e95-acc3-116261551091"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.253113 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/804394f9-a48a-4e95-acc3-116261551091-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "804394f9-a48a-4e95-acc3-116261551091" (UID: "804394f9-a48a-4e95-acc3-116261551091"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.319527 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v89v4\" (UniqueName: \"kubernetes.io/projected/804394f9-a48a-4e95-acc3-116261551091-kube-api-access-v89v4\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.319572 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/804394f9-a48a-4e95-acc3-116261551091-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.319585 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/804394f9-a48a-4e95-acc3-116261551091-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.496530 4894 generic.go:334] "Generic (PLEG): container finished" podID="804394f9-a48a-4e95-acc3-116261551091" containerID="4c18fb89fb255d15b9bf0fba6f95c842299b7c92e911ad69bf975428c012a38a" exitCode=0 Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.496573 4894 generic.go:334] "Generic (PLEG): container finished" podID="804394f9-a48a-4e95-acc3-116261551091" containerID="07ab85febb0174b07422933bf634288d5ebc79aee758a336123a62e63063c133" exitCode=143 Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.496592 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.496653 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"804394f9-a48a-4e95-acc3-116261551091","Type":"ContainerDied","Data":"4c18fb89fb255d15b9bf0fba6f95c842299b7c92e911ad69bf975428c012a38a"} Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.496707 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"804394f9-a48a-4e95-acc3-116261551091","Type":"ContainerDied","Data":"07ab85febb0174b07422933bf634288d5ebc79aee758a336123a62e63063c133"} Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.496721 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"804394f9-a48a-4e95-acc3-116261551091","Type":"ContainerDied","Data":"15bf1587a4d0691b9528084b312b389f0f6f0af31fc1e278d1f7865d82091dc7"} Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.496738 4894 scope.go:117] "RemoveContainer" containerID="4c18fb89fb255d15b9bf0fba6f95c842299b7c92e911ad69bf975428c012a38a" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.533530 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.542457 4894 scope.go:117] "RemoveContainer" containerID="07ab85febb0174b07422933bf634288d5ebc79aee758a336123a62e63063c133" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.547738 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.565277 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 09 15:51:56 crc kubenswrapper[4894]: E1209 15:51:56.567271 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="804394f9-a48a-4e95-acc3-116261551091" containerName="nova-metadata-metadata" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.567397 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="804394f9-a48a-4e95-acc3-116261551091" containerName="nova-metadata-metadata" Dec 09 15:51:56 crc kubenswrapper[4894]: E1209 15:51:56.567500 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="804394f9-a48a-4e95-acc3-116261551091" containerName="nova-metadata-log" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.567556 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="804394f9-a48a-4e95-acc3-116261551091" containerName="nova-metadata-log" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.567832 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="804394f9-a48a-4e95-acc3-116261551091" containerName="nova-metadata-log" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.567904 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="804394f9-a48a-4e95-acc3-116261551091" containerName="nova-metadata-metadata" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.570020 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.572761 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.573222 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.586444 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.594759 4894 scope.go:117] "RemoveContainer" containerID="4c18fb89fb255d15b9bf0fba6f95c842299b7c92e911ad69bf975428c012a38a" Dec 09 15:51:56 crc kubenswrapper[4894]: E1209 15:51:56.596326 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c18fb89fb255d15b9bf0fba6f95c842299b7c92e911ad69bf975428c012a38a\": container with ID starting with 4c18fb89fb255d15b9bf0fba6f95c842299b7c92e911ad69bf975428c012a38a not found: ID does not exist" containerID="4c18fb89fb255d15b9bf0fba6f95c842299b7c92e911ad69bf975428c012a38a" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.596365 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c18fb89fb255d15b9bf0fba6f95c842299b7c92e911ad69bf975428c012a38a"} err="failed to get container status \"4c18fb89fb255d15b9bf0fba6f95c842299b7c92e911ad69bf975428c012a38a\": rpc error: code = NotFound desc = could not find container \"4c18fb89fb255d15b9bf0fba6f95c842299b7c92e911ad69bf975428c012a38a\": container with ID starting with 4c18fb89fb255d15b9bf0fba6f95c842299b7c92e911ad69bf975428c012a38a not found: ID does not exist" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.596390 4894 scope.go:117] "RemoveContainer" containerID="07ab85febb0174b07422933bf634288d5ebc79aee758a336123a62e63063c133" Dec 09 15:51:56 crc kubenswrapper[4894]: E1209 15:51:56.598892 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"07ab85febb0174b07422933bf634288d5ebc79aee758a336123a62e63063c133\": container with ID starting with 07ab85febb0174b07422933bf634288d5ebc79aee758a336123a62e63063c133 not found: ID does not exist" containerID="07ab85febb0174b07422933bf634288d5ebc79aee758a336123a62e63063c133" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.598917 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07ab85febb0174b07422933bf634288d5ebc79aee758a336123a62e63063c133"} err="failed to get container status \"07ab85febb0174b07422933bf634288d5ebc79aee758a336123a62e63063c133\": rpc error: code = NotFound desc = could not find container \"07ab85febb0174b07422933bf634288d5ebc79aee758a336123a62e63063c133\": container with ID starting with 07ab85febb0174b07422933bf634288d5ebc79aee758a336123a62e63063c133 not found: ID does not exist" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.598934 4894 scope.go:117] "RemoveContainer" containerID="4c18fb89fb255d15b9bf0fba6f95c842299b7c92e911ad69bf975428c012a38a" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.600627 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c18fb89fb255d15b9bf0fba6f95c842299b7c92e911ad69bf975428c012a38a"} err="failed to get container status \"4c18fb89fb255d15b9bf0fba6f95c842299b7c92e911ad69bf975428c012a38a\": rpc error: code = NotFound desc = could not find container \"4c18fb89fb255d15b9bf0fba6f95c842299b7c92e911ad69bf975428c012a38a\": container with ID starting with 4c18fb89fb255d15b9bf0fba6f95c842299b7c92e911ad69bf975428c012a38a not found: ID does not exist" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.600671 4894 scope.go:117] "RemoveContainer" containerID="07ab85febb0174b07422933bf634288d5ebc79aee758a336123a62e63063c133" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.608827 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"07ab85febb0174b07422933bf634288d5ebc79aee758a336123a62e63063c133"} err="failed to get container status \"07ab85febb0174b07422933bf634288d5ebc79aee758a336123a62e63063c133\": rpc error: code = NotFound desc = could not find container \"07ab85febb0174b07422933bf634288d5ebc79aee758a336123a62e63063c133\": container with ID starting with 07ab85febb0174b07422933bf634288d5ebc79aee758a336123a62e63063c133 not found: ID does not exist" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.627094 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\") " pod="openstack/nova-metadata-0" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.627170 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-config-data\") pod \"nova-metadata-0\" (UID: \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\") " pod="openstack/nova-metadata-0" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.627207 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-logs\") pod \"nova-metadata-0\" (UID: \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\") " pod="openstack/nova-metadata-0" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.627237 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gw9wb\" (UniqueName: \"kubernetes.io/projected/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-kube-api-access-gw9wb\") pod \"nova-metadata-0\" (UID: \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\") " pod="openstack/nova-metadata-0" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.627489 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\") " pod="openstack/nova-metadata-0" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.729253 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\") " pod="openstack/nova-metadata-0" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.729361 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\") " pod="openstack/nova-metadata-0" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.729404 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-config-data\") pod \"nova-metadata-0\" (UID: \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\") " pod="openstack/nova-metadata-0" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.729429 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-logs\") pod \"nova-metadata-0\" (UID: \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\") " pod="openstack/nova-metadata-0" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.729459 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gw9wb\" (UniqueName: \"kubernetes.io/projected/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-kube-api-access-gw9wb\") pod \"nova-metadata-0\" (UID: \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\") " pod="openstack/nova-metadata-0" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.729955 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-logs\") pod \"nova-metadata-0\" (UID: \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\") " pod="openstack/nova-metadata-0" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.734350 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\") " pod="openstack/nova-metadata-0" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.734623 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-config-data\") pod \"nova-metadata-0\" (UID: \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\") " pod="openstack/nova-metadata-0" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.736393 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\") " pod="openstack/nova-metadata-0" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.748509 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gw9wb\" (UniqueName: \"kubernetes.io/projected/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-kube-api-access-gw9wb\") pod \"nova-metadata-0\" (UID: \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\") " pod="openstack/nova-metadata-0" Dec 09 15:51:56 crc kubenswrapper[4894]: I1209 15:51:56.901488 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 15:51:57 crc kubenswrapper[4894]: I1209 15:51:57.384109 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 15:51:57 crc kubenswrapper[4894]: I1209 15:51:57.506621 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"42cddbce-4418-4bd5-9012-ac8b7e6bd26a","Type":"ContainerStarted","Data":"8a46f2057ec37f88804434c59e85bbfa05463a894237c36f633533d689b836a8"} Dec 09 15:51:57 crc kubenswrapper[4894]: I1209 15:51:57.508750 4894 generic.go:334] "Generic (PLEG): container finished" podID="e2f0b872-8eba-4309-94c8-4dfe893cfc3c" containerID="389f6a7420333dcd236202ca16cd594b4e9bb3dd3e8434999e9784a6d984c7c9" exitCode=0 Dec 09 15:51:57 crc kubenswrapper[4894]: I1209 15:51:57.508812 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-w26bv" event={"ID":"e2f0b872-8eba-4309-94c8-4dfe893cfc3c","Type":"ContainerDied","Data":"389f6a7420333dcd236202ca16cd594b4e9bb3dd3e8434999e9784a6d984c7c9"} Dec 09 15:51:57 crc kubenswrapper[4894]: I1209 15:51:57.535312 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:51:57 crc kubenswrapper[4894]: I1209 15:51:57.765357 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 09 15:51:57 crc kubenswrapper[4894]: I1209 15:51:57.765750 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 09 15:51:57 crc kubenswrapper[4894]: I1209 15:51:57.807023 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 09 15:51:57 crc kubenswrapper[4894]: I1209 15:51:57.890732 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 15:51:57 crc kubenswrapper[4894]: I1209 15:51:57.890787 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 15:51:58 crc kubenswrapper[4894]: I1209 15:51:58.125785 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="804394f9-a48a-4e95-acc3-116261551091" path="/var/lib/kubelet/pods/804394f9-a48a-4e95-acc3-116261551091/volumes" Dec 09 15:51:58 crc kubenswrapper[4894]: I1209 15:51:58.461186 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 09 15:51:58 crc kubenswrapper[4894]: I1209 15:51:58.530124 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"42cddbce-4418-4bd5-9012-ac8b7e6bd26a","Type":"ContainerStarted","Data":"68db72b024449a5cf0cc8880a3cabd6a93e9c5b7016c37ec22575ab627556bf1"} Dec 09 15:51:58 crc kubenswrapper[4894]: I1209 15:51:58.530188 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"42cddbce-4418-4bd5-9012-ac8b7e6bd26a","Type":"ContainerStarted","Data":"a1294d691b406848d0155ce9534a1770cf2cc113a3d429ad86b19f9240c27d5e"} Dec 09 15:51:58 crc kubenswrapper[4894]: I1209 15:51:58.561402 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.56138123 podStartE2EDuration="2.56138123s" podCreationTimestamp="2025-12-09 15:51:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:51:58.550855054 +0000 UTC m=+1212.870065743" watchObservedRunningTime="2025-12-09 15:51:58.56138123 +0000 UTC m=+1212.880591899" Dec 09 15:51:58 crc kubenswrapper[4894]: I1209 15:51:58.578829 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 09 15:51:58 crc kubenswrapper[4894]: I1209 15:51:58.977199 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9418641c-296a-4d30-a02a-52c150f99cd2" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.179:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 15:51:58 crc kubenswrapper[4894]: I1209 15:51:58.977320 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="9418641c-296a-4d30-a02a-52c150f99cd2" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.179:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.115198 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-w26bv" Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.187797 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-config-data\") pod \"e2f0b872-8eba-4309-94c8-4dfe893cfc3c\" (UID: \"e2f0b872-8eba-4309-94c8-4dfe893cfc3c\") " Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.187910 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-combined-ca-bundle\") pod \"e2f0b872-8eba-4309-94c8-4dfe893cfc3c\" (UID: \"e2f0b872-8eba-4309-94c8-4dfe893cfc3c\") " Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.188023 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-scripts\") pod \"e2f0b872-8eba-4309-94c8-4dfe893cfc3c\" (UID: \"e2f0b872-8eba-4309-94c8-4dfe893cfc3c\") " Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.188050 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jx6dg\" (UniqueName: \"kubernetes.io/projected/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-kube-api-access-jx6dg\") pod \"e2f0b872-8eba-4309-94c8-4dfe893cfc3c\" (UID: \"e2f0b872-8eba-4309-94c8-4dfe893cfc3c\") " Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.193568 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-scripts" (OuterVolumeSpecName: "scripts") pod "e2f0b872-8eba-4309-94c8-4dfe893cfc3c" (UID: "e2f0b872-8eba-4309-94c8-4dfe893cfc3c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.196988 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-kube-api-access-jx6dg" (OuterVolumeSpecName: "kube-api-access-jx6dg") pod "e2f0b872-8eba-4309-94c8-4dfe893cfc3c" (UID: "e2f0b872-8eba-4309-94c8-4dfe893cfc3c"). InnerVolumeSpecName "kube-api-access-jx6dg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.222594 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e2f0b872-8eba-4309-94c8-4dfe893cfc3c" (UID: "e2f0b872-8eba-4309-94c8-4dfe893cfc3c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.239961 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-config-data" (OuterVolumeSpecName: "config-data") pod "e2f0b872-8eba-4309-94c8-4dfe893cfc3c" (UID: "e2f0b872-8eba-4309-94c8-4dfe893cfc3c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.290899 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.290942 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.290956 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.290972 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jx6dg\" (UniqueName: \"kubernetes.io/projected/e2f0b872-8eba-4309-94c8-4dfe893cfc3c-kube-api-access-jx6dg\") on node \"crc\" DevicePath \"\"" Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.538693 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-w26bv" Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.538614 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-w26bv" event={"ID":"e2f0b872-8eba-4309-94c8-4dfe893cfc3c","Type":"ContainerDied","Data":"0640508dec8da6be94969fbbef9bffa9766bd27992ad1a305924706a40a987ff"} Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.538783 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0640508dec8da6be94969fbbef9bffa9766bd27992ad1a305924706a40a987ff" Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.540201 4894 generic.go:334] "Generic (PLEG): container finished" podID="0f261fad-179c-4f25-bc42-44691e48be50" containerID="29e9cc197d569126cf8be32f02028738bb59d1c4167b6af1d9d999aecd711a2e" exitCode=0 Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.540288 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-scszh" event={"ID":"0f261fad-179c-4f25-bc42-44691e48be50","Type":"ContainerDied","Data":"29e9cc197d569126cf8be32f02028738bb59d1c4167b6af1d9d999aecd711a2e"} Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.724028 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.724596 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9418641c-296a-4d30-a02a-52c150f99cd2" containerName="nova-api-log" containerID="cri-o://ec90db2fbe4702580bec2f02502f736637c41093f6c7f277c82047e5b880698f" gracePeriod=30 Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.724733 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9418641c-296a-4d30-a02a-52c150f99cd2" containerName="nova-api-api" containerID="cri-o://8687a54d3c71b9460c2c6bfe89afe97dd96d5b812daf0a197b0df3633e5b6c66" gracePeriod=30 Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.754237 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 15:51:59 crc kubenswrapper[4894]: I1209 15:51:59.801500 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 15:52:00 crc kubenswrapper[4894]: I1209 15:52:00.549742 4894 generic.go:334] "Generic (PLEG): container finished" podID="9418641c-296a-4d30-a02a-52c150f99cd2" containerID="ec90db2fbe4702580bec2f02502f736637c41093f6c7f277c82047e5b880698f" exitCode=143 Dec 09 15:52:00 crc kubenswrapper[4894]: I1209 15:52:00.549872 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9418641c-296a-4d30-a02a-52c150f99cd2","Type":"ContainerDied","Data":"ec90db2fbe4702580bec2f02502f736637c41093f6c7f277c82047e5b880698f"} Dec 09 15:52:00 crc kubenswrapper[4894]: I1209 15:52:00.550054 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="42cddbce-4418-4bd5-9012-ac8b7e6bd26a" containerName="nova-metadata-log" containerID="cri-o://a1294d691b406848d0155ce9534a1770cf2cc113a3d429ad86b19f9240c27d5e" gracePeriod=30 Dec 09 15:52:00 crc kubenswrapper[4894]: I1209 15:52:00.550677 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="42cddbce-4418-4bd5-9012-ac8b7e6bd26a" containerName="nova-metadata-metadata" containerID="cri-o://68db72b024449a5cf0cc8880a3cabd6a93e9c5b7016c37ec22575ab627556bf1" gracePeriod=30 Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.000496 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-scszh" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.127211 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f261fad-179c-4f25-bc42-44691e48be50-combined-ca-bundle\") pod \"0f261fad-179c-4f25-bc42-44691e48be50\" (UID: \"0f261fad-179c-4f25-bc42-44691e48be50\") " Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.127302 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jk65l\" (UniqueName: \"kubernetes.io/projected/0f261fad-179c-4f25-bc42-44691e48be50-kube-api-access-jk65l\") pod \"0f261fad-179c-4f25-bc42-44691e48be50\" (UID: \"0f261fad-179c-4f25-bc42-44691e48be50\") " Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.127361 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f261fad-179c-4f25-bc42-44691e48be50-config-data\") pod \"0f261fad-179c-4f25-bc42-44691e48be50\" (UID: \"0f261fad-179c-4f25-bc42-44691e48be50\") " Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.127389 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f261fad-179c-4f25-bc42-44691e48be50-scripts\") pod \"0f261fad-179c-4f25-bc42-44691e48be50\" (UID: \"0f261fad-179c-4f25-bc42-44691e48be50\") " Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.134666 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f261fad-179c-4f25-bc42-44691e48be50-scripts" (OuterVolumeSpecName: "scripts") pod "0f261fad-179c-4f25-bc42-44691e48be50" (UID: "0f261fad-179c-4f25-bc42-44691e48be50"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.134959 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0f261fad-179c-4f25-bc42-44691e48be50-kube-api-access-jk65l" (OuterVolumeSpecName: "kube-api-access-jk65l") pod "0f261fad-179c-4f25-bc42-44691e48be50" (UID: "0f261fad-179c-4f25-bc42-44691e48be50"). InnerVolumeSpecName "kube-api-access-jk65l". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.145027 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.169213 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f261fad-179c-4f25-bc42-44691e48be50-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0f261fad-179c-4f25-bc42-44691e48be50" (UID: "0f261fad-179c-4f25-bc42-44691e48be50"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.180625 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0f261fad-179c-4f25-bc42-44691e48be50-config-data" (OuterVolumeSpecName: "config-data") pod "0f261fad-179c-4f25-bc42-44691e48be50" (UID: "0f261fad-179c-4f25-bc42-44691e48be50"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.228616 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gw9wb\" (UniqueName: \"kubernetes.io/projected/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-kube-api-access-gw9wb\") pod \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\" (UID: \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\") " Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.228716 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-nova-metadata-tls-certs\") pod \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\" (UID: \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\") " Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.228840 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-config-data\") pod \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\" (UID: \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\") " Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.228930 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-logs\") pod \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\" (UID: \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\") " Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.228954 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-combined-ca-bundle\") pod \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\" (UID: \"42cddbce-4418-4bd5-9012-ac8b7e6bd26a\") " Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.229568 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0f261fad-179c-4f25-bc42-44691e48be50-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.229586 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0f261fad-179c-4f25-bc42-44691e48be50-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.229622 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jk65l\" (UniqueName: \"kubernetes.io/projected/0f261fad-179c-4f25-bc42-44691e48be50-kube-api-access-jk65l\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.229631 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0f261fad-179c-4f25-bc42-44691e48be50-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.231821 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-logs" (OuterVolumeSpecName: "logs") pod "42cddbce-4418-4bd5-9012-ac8b7e6bd26a" (UID: "42cddbce-4418-4bd5-9012-ac8b7e6bd26a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.233628 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-kube-api-access-gw9wb" (OuterVolumeSpecName: "kube-api-access-gw9wb") pod "42cddbce-4418-4bd5-9012-ac8b7e6bd26a" (UID: "42cddbce-4418-4bd5-9012-ac8b7e6bd26a"). InnerVolumeSpecName "kube-api-access-gw9wb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.265303 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "42cddbce-4418-4bd5-9012-ac8b7e6bd26a" (UID: "42cddbce-4418-4bd5-9012-ac8b7e6bd26a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.267250 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.267520 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="48d401c8-b930-410e-a313-80399a664641" containerName="kube-state-metrics" containerID="cri-o://94f37c74641c85bdc8795734b07ff70ec2f8ac7e3c25edf2aac47eace970074d" gracePeriod=30 Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.277373 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-config-data" (OuterVolumeSpecName: "config-data") pod "42cddbce-4418-4bd5-9012-ac8b7e6bd26a" (UID: "42cddbce-4418-4bd5-9012-ac8b7e6bd26a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.303078 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "42cddbce-4418-4bd5-9012-ac8b7e6bd26a" (UID: "42cddbce-4418-4bd5-9012-ac8b7e6bd26a"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.331549 4894 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.331595 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.331609 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-logs\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.331622 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.331655 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gw9wb\" (UniqueName: \"kubernetes.io/projected/42cddbce-4418-4bd5-9012-ac8b7e6bd26a-kube-api-access-gw9wb\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.574343 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-scszh" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.574343 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-scszh" event={"ID":"0f261fad-179c-4f25-bc42-44691e48be50","Type":"ContainerDied","Data":"59dd9db98efca6ebed794359db462ea823df28e1411babd70f8d682850b5da2f"} Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.574799 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="59dd9db98efca6ebed794359db462ea823df28e1411babd70f8d682850b5da2f" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.582345 4894 generic.go:334] "Generic (PLEG): container finished" podID="42cddbce-4418-4bd5-9012-ac8b7e6bd26a" containerID="68db72b024449a5cf0cc8880a3cabd6a93e9c5b7016c37ec22575ab627556bf1" exitCode=0 Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.582377 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.582386 4894 generic.go:334] "Generic (PLEG): container finished" podID="42cddbce-4418-4bd5-9012-ac8b7e6bd26a" containerID="a1294d691b406848d0155ce9534a1770cf2cc113a3d429ad86b19f9240c27d5e" exitCode=143 Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.582389 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"42cddbce-4418-4bd5-9012-ac8b7e6bd26a","Type":"ContainerDied","Data":"68db72b024449a5cf0cc8880a3cabd6a93e9c5b7016c37ec22575ab627556bf1"} Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.582430 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"42cddbce-4418-4bd5-9012-ac8b7e6bd26a","Type":"ContainerDied","Data":"a1294d691b406848d0155ce9534a1770cf2cc113a3d429ad86b19f9240c27d5e"} Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.582446 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"42cddbce-4418-4bd5-9012-ac8b7e6bd26a","Type":"ContainerDied","Data":"8a46f2057ec37f88804434c59e85bbfa05463a894237c36f633533d689b836a8"} Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.582460 4894 scope.go:117] "RemoveContainer" containerID="68db72b024449a5cf0cc8880a3cabd6a93e9c5b7016c37ec22575ab627556bf1" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.585409 4894 generic.go:334] "Generic (PLEG): container finished" podID="48d401c8-b930-410e-a313-80399a664641" containerID="94f37c74641c85bdc8795734b07ff70ec2f8ac7e3c25edf2aac47eace970074d" exitCode=2 Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.585523 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"48d401c8-b930-410e-a313-80399a664641","Type":"ContainerDied","Data":"94f37c74641c85bdc8795734b07ff70ec2f8ac7e3c25edf2aac47eace970074d"} Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.585572 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="8cefa8f8-0c44-4827-af46-671cea04dd97" containerName="nova-scheduler-scheduler" containerID="cri-o://f2556776659526f969cb1e90d6261ee3f6e0696cae9699946f744b0fac60b09b" gracePeriod=30 Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.662855 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 09 15:52:01 crc kubenswrapper[4894]: E1209 15:52:01.663323 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0f261fad-179c-4f25-bc42-44691e48be50" containerName="nova-cell1-conductor-db-sync" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.663340 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="0f261fad-179c-4f25-bc42-44691e48be50" containerName="nova-cell1-conductor-db-sync" Dec 09 15:52:01 crc kubenswrapper[4894]: E1209 15:52:01.663358 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42cddbce-4418-4bd5-9012-ac8b7e6bd26a" containerName="nova-metadata-log" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.663366 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="42cddbce-4418-4bd5-9012-ac8b7e6bd26a" containerName="nova-metadata-log" Dec 09 15:52:01 crc kubenswrapper[4894]: E1209 15:52:01.663380 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="42cddbce-4418-4bd5-9012-ac8b7e6bd26a" containerName="nova-metadata-metadata" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.663388 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="42cddbce-4418-4bd5-9012-ac8b7e6bd26a" containerName="nova-metadata-metadata" Dec 09 15:52:01 crc kubenswrapper[4894]: E1209 15:52:01.663405 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e2f0b872-8eba-4309-94c8-4dfe893cfc3c" containerName="nova-manage" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.663414 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e2f0b872-8eba-4309-94c8-4dfe893cfc3c" containerName="nova-manage" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.663688 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e2f0b872-8eba-4309-94c8-4dfe893cfc3c" containerName="nova-manage" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.663705 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="42cddbce-4418-4bd5-9012-ac8b7e6bd26a" containerName="nova-metadata-log" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.663719 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="42cddbce-4418-4bd5-9012-ac8b7e6bd26a" containerName="nova-metadata-metadata" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.663743 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="0f261fad-179c-4f25-bc42-44691e48be50" containerName="nova-cell1-conductor-db-sync" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.664169 4894 scope.go:117] "RemoveContainer" containerID="a1294d691b406848d0155ce9534a1770cf2cc113a3d429ad86b19f9240c27d5e" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.664457 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.674318 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.722390 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.741262 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.749470 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00ff6ff2-b452-45ae-8f9d-314363bcb126-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"00ff6ff2-b452-45ae-8f9d-314363bcb126\") " pod="openstack/nova-cell1-conductor-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.749575 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00ff6ff2-b452-45ae-8f9d-314363bcb126-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"00ff6ff2-b452-45ae-8f9d-314363bcb126\") " pod="openstack/nova-cell1-conductor-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.749633 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vcbp9\" (UniqueName: \"kubernetes.io/projected/00ff6ff2-b452-45ae-8f9d-314363bcb126-kube-api-access-vcbp9\") pod \"nova-cell1-conductor-0\" (UID: \"00ff6ff2-b452-45ae-8f9d-314363bcb126\") " pod="openstack/nova-cell1-conductor-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.756239 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.762786 4894 scope.go:117] "RemoveContainer" containerID="68db72b024449a5cf0cc8880a3cabd6a93e9c5b7016c37ec22575ab627556bf1" Dec 09 15:52:01 crc kubenswrapper[4894]: E1209 15:52:01.763257 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"68db72b024449a5cf0cc8880a3cabd6a93e9c5b7016c37ec22575ab627556bf1\": container with ID starting with 68db72b024449a5cf0cc8880a3cabd6a93e9c5b7016c37ec22575ab627556bf1 not found: ID does not exist" containerID="68db72b024449a5cf0cc8880a3cabd6a93e9c5b7016c37ec22575ab627556bf1" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.763285 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68db72b024449a5cf0cc8880a3cabd6a93e9c5b7016c37ec22575ab627556bf1"} err="failed to get container status \"68db72b024449a5cf0cc8880a3cabd6a93e9c5b7016c37ec22575ab627556bf1\": rpc error: code = NotFound desc = could not find container \"68db72b024449a5cf0cc8880a3cabd6a93e9c5b7016c37ec22575ab627556bf1\": container with ID starting with 68db72b024449a5cf0cc8880a3cabd6a93e9c5b7016c37ec22575ab627556bf1 not found: ID does not exist" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.763305 4894 scope.go:117] "RemoveContainer" containerID="a1294d691b406848d0155ce9534a1770cf2cc113a3d429ad86b19f9240c27d5e" Dec 09 15:52:01 crc kubenswrapper[4894]: E1209 15:52:01.763536 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1294d691b406848d0155ce9534a1770cf2cc113a3d429ad86b19f9240c27d5e\": container with ID starting with a1294d691b406848d0155ce9534a1770cf2cc113a3d429ad86b19f9240c27d5e not found: ID does not exist" containerID="a1294d691b406848d0155ce9534a1770cf2cc113a3d429ad86b19f9240c27d5e" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.763558 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1294d691b406848d0155ce9534a1770cf2cc113a3d429ad86b19f9240c27d5e"} err="failed to get container status \"a1294d691b406848d0155ce9534a1770cf2cc113a3d429ad86b19f9240c27d5e\": rpc error: code = NotFound desc = could not find container \"a1294d691b406848d0155ce9534a1770cf2cc113a3d429ad86b19f9240c27d5e\": container with ID starting with a1294d691b406848d0155ce9534a1770cf2cc113a3d429ad86b19f9240c27d5e not found: ID does not exist" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.763577 4894 scope.go:117] "RemoveContainer" containerID="68db72b024449a5cf0cc8880a3cabd6a93e9c5b7016c37ec22575ab627556bf1" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.763893 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"68db72b024449a5cf0cc8880a3cabd6a93e9c5b7016c37ec22575ab627556bf1"} err="failed to get container status \"68db72b024449a5cf0cc8880a3cabd6a93e9c5b7016c37ec22575ab627556bf1\": rpc error: code = NotFound desc = could not find container \"68db72b024449a5cf0cc8880a3cabd6a93e9c5b7016c37ec22575ab627556bf1\": container with ID starting with 68db72b024449a5cf0cc8880a3cabd6a93e9c5b7016c37ec22575ab627556bf1 not found: ID does not exist" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.763912 4894 scope.go:117] "RemoveContainer" containerID="a1294d691b406848d0155ce9534a1770cf2cc113a3d429ad86b19f9240c27d5e" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.764875 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1294d691b406848d0155ce9534a1770cf2cc113a3d429ad86b19f9240c27d5e"} err="failed to get container status \"a1294d691b406848d0155ce9534a1770cf2cc113a3d429ad86b19f9240c27d5e\": rpc error: code = NotFound desc = could not find container \"a1294d691b406848d0155ce9534a1770cf2cc113a3d429ad86b19f9240c27d5e\": container with ID starting with a1294d691b406848d0155ce9534a1770cf2cc113a3d429ad86b19f9240c27d5e not found: ID does not exist" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.791005 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.792857 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.795494 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.796078 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.836707 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.851703 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vcbp9\" (UniqueName: \"kubernetes.io/projected/00ff6ff2-b452-45ae-8f9d-314363bcb126-kube-api-access-vcbp9\") pod \"nova-cell1-conductor-0\" (UID: \"00ff6ff2-b452-45ae-8f9d-314363bcb126\") " pod="openstack/nova-cell1-conductor-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.852259 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-config-data\") pod \"nova-metadata-0\" (UID: \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\") " pod="openstack/nova-metadata-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.852433 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00ff6ff2-b452-45ae-8f9d-314363bcb126-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"00ff6ff2-b452-45ae-8f9d-314363bcb126\") " pod="openstack/nova-cell1-conductor-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.852565 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9dnw2\" (UniqueName: \"kubernetes.io/projected/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-kube-api-access-9dnw2\") pod \"nova-metadata-0\" (UID: \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\") " pod="openstack/nova-metadata-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.852713 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\") " pod="openstack/nova-metadata-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.852868 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-logs\") pod \"nova-metadata-0\" (UID: \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\") " pod="openstack/nova-metadata-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.852996 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\") " pod="openstack/nova-metadata-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.853113 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00ff6ff2-b452-45ae-8f9d-314363bcb126-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"00ff6ff2-b452-45ae-8f9d-314363bcb126\") " pod="openstack/nova-cell1-conductor-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.864476 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/00ff6ff2-b452-45ae-8f9d-314363bcb126-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"00ff6ff2-b452-45ae-8f9d-314363bcb126\") " pod="openstack/nova-cell1-conductor-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.868346 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/00ff6ff2-b452-45ae-8f9d-314363bcb126-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"00ff6ff2-b452-45ae-8f9d-314363bcb126\") " pod="openstack/nova-cell1-conductor-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.877358 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vcbp9\" (UniqueName: \"kubernetes.io/projected/00ff6ff2-b452-45ae-8f9d-314363bcb126-kube-api-access-vcbp9\") pod \"nova-cell1-conductor-0\" (UID: \"00ff6ff2-b452-45ae-8f9d-314363bcb126\") " pod="openstack/nova-cell1-conductor-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.954400 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-config-data\") pod \"nova-metadata-0\" (UID: \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\") " pod="openstack/nova-metadata-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.954533 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9dnw2\" (UniqueName: \"kubernetes.io/projected/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-kube-api-access-9dnw2\") pod \"nova-metadata-0\" (UID: \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\") " pod="openstack/nova-metadata-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.954560 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\") " pod="openstack/nova-metadata-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.954623 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-logs\") pod \"nova-metadata-0\" (UID: \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\") " pod="openstack/nova-metadata-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.954665 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\") " pod="openstack/nova-metadata-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.955659 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-logs\") pod \"nova-metadata-0\" (UID: \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\") " pod="openstack/nova-metadata-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.960321 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\") " pod="openstack/nova-metadata-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.966932 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.986563 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9dnw2\" (UniqueName: \"kubernetes.io/projected/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-kube-api-access-9dnw2\") pod \"nova-metadata-0\" (UID: \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\") " pod="openstack/nova-metadata-0" Dec 09 15:52:01 crc kubenswrapper[4894]: I1209 15:52:01.999466 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\") " pod="openstack/nova-metadata-0" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.004986 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-config-data\") pod \"nova-metadata-0\" (UID: \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\") " pod="openstack/nova-metadata-0" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.055887 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gk2n5\" (UniqueName: \"kubernetes.io/projected/48d401c8-b930-410e-a313-80399a664641-kube-api-access-gk2n5\") pod \"48d401c8-b930-410e-a313-80399a664641\" (UID: \"48d401c8-b930-410e-a313-80399a664641\") " Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.061829 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48d401c8-b930-410e-a313-80399a664641-kube-api-access-gk2n5" (OuterVolumeSpecName: "kube-api-access-gk2n5") pod "48d401c8-b930-410e-a313-80399a664641" (UID: "48d401c8-b930-410e-a313-80399a664641"). InnerVolumeSpecName "kube-api-access-gk2n5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.118245 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="42cddbce-4418-4bd5-9012-ac8b7e6bd26a" path="/var/lib/kubelet/pods/42cddbce-4418-4bd5-9012-ac8b7e6bd26a/volumes" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.159675 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gk2n5\" (UniqueName: \"kubernetes.io/projected/48d401c8-b930-410e-a313-80399a664641-kube-api-access-gk2n5\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.164712 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.285711 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.606126 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"48d401c8-b930-410e-a313-80399a664641","Type":"ContainerDied","Data":"afd9f3c9cc738ec79ce7df87c0ecd6b32a700e7b65cde148fe85b718b531f086"} Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.606442 4894 scope.go:117] "RemoveContainer" containerID="94f37c74641c85bdc8795734b07ff70ec2f8ac7e3c25edf2aac47eace970074d" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.606591 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 15:52:02 crc kubenswrapper[4894]: W1209 15:52:02.610359 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod00ff6ff2_b452_45ae_8f9d_314363bcb126.slice/crio-f1da57090a0165790834d0b1a3a5fab5c3717555960a6d99189b8c7217b1a0ca WatchSource:0}: Error finding container f1da57090a0165790834d0b1a3a5fab5c3717555960a6d99189b8c7217b1a0ca: Status 404 returned error can't find the container with id f1da57090a0165790834d0b1a3a5fab5c3717555960a6d99189b8c7217b1a0ca Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.613139 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.652876 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.657813 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.667684 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 15:52:02 crc kubenswrapper[4894]: E1209 15:52:02.668113 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48d401c8-b930-410e-a313-80399a664641" containerName="kube-state-metrics" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.668128 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="48d401c8-b930-410e-a313-80399a664641" containerName="kube-state-metrics" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.668346 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="48d401c8-b930-410e-a313-80399a664641" containerName="kube-state-metrics" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.669223 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.673265 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.673579 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.675032 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.756357 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 15:52:02 crc kubenswrapper[4894]: W1209 15:52:02.767759 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode97b390d_e40e_45d2_8eb7_c4ff7205e5fb.slice/crio-62493d8e3b705c3e5b04425414000d458b9741f791db79fa72c7af19236476e3 WatchSource:0}: Error finding container 62493d8e3b705c3e5b04425414000d458b9741f791db79fa72c7af19236476e3: Status 404 returned error can't find the container with id 62493d8e3b705c3e5b04425414000d458b9741f791db79fa72c7af19236476e3 Dec 09 15:52:02 crc kubenswrapper[4894]: E1209 15:52:02.767837 4894 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f2556776659526f969cb1e90d6261ee3f6e0696cae9699946f744b0fac60b09b is running failed: container process not found" containerID="f2556776659526f969cb1e90d6261ee3f6e0696cae9699946f744b0fac60b09b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 15:52:02 crc kubenswrapper[4894]: E1209 15:52:02.768530 4894 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f2556776659526f969cb1e90d6261ee3f6e0696cae9699946f744b0fac60b09b is running failed: container process not found" containerID="f2556776659526f969cb1e90d6261ee3f6e0696cae9699946f744b0fac60b09b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 15:52:02 crc kubenswrapper[4894]: E1209 15:52:02.768824 4894 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f2556776659526f969cb1e90d6261ee3f6e0696cae9699946f744b0fac60b09b is running failed: container process not found" containerID="f2556776659526f969cb1e90d6261ee3f6e0696cae9699946f744b0fac60b09b" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 15:52:02 crc kubenswrapper[4894]: E1209 15:52:02.768852 4894 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of f2556776659526f969cb1e90d6261ee3f6e0696cae9699946f744b0fac60b09b is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="8cefa8f8-0c44-4827-af46-671cea04dd97" containerName="nova-scheduler-scheduler" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.776626 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/62f0c278-8dcb-4e9d-8844-98f8b946edd8-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"62f0c278-8dcb-4e9d-8844-98f8b946edd8\") " pod="openstack/kube-state-metrics-0" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.776939 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62f0c278-8dcb-4e9d-8844-98f8b946edd8-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"62f0c278-8dcb-4e9d-8844-98f8b946edd8\") " pod="openstack/kube-state-metrics-0" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.777021 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/62f0c278-8dcb-4e9d-8844-98f8b946edd8-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"62f0c278-8dcb-4e9d-8844-98f8b946edd8\") " pod="openstack/kube-state-metrics-0" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.777175 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6qk52\" (UniqueName: \"kubernetes.io/projected/62f0c278-8dcb-4e9d-8844-98f8b946edd8-kube-api-access-6qk52\") pod \"kube-state-metrics-0\" (UID: \"62f0c278-8dcb-4e9d-8844-98f8b946edd8\") " pod="openstack/kube-state-metrics-0" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.808908 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.810092 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7fc523c0-a7fc-4689-a3ed-e43288589249" containerName="ceilometer-central-agent" containerID="cri-o://67eccfb195452e20fd8a6e68c0898294288d59b21488e44bb1ec66b34cf286f6" gracePeriod=30 Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.810174 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7fc523c0-a7fc-4689-a3ed-e43288589249" containerName="sg-core" containerID="cri-o://860d30ad535d6172272c81311f33b40c1de0b026651c3282663f5518b92bb13a" gracePeriod=30 Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.810205 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7fc523c0-a7fc-4689-a3ed-e43288589249" containerName="ceilometer-notification-agent" containerID="cri-o://df527646f6cc5b4c20571dd4d6a938bc0029167739028aac8ab7f819a664e895" gracePeriod=30 Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.810350 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="7fc523c0-a7fc-4689-a3ed-e43288589249" containerName="proxy-httpd" containerID="cri-o://7460daec013ac6ac4932f395ede5e82849994c5daabe2e8383db32ee12fc51a9" gracePeriod=30 Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.879070 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62f0c278-8dcb-4e9d-8844-98f8b946edd8-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"62f0c278-8dcb-4e9d-8844-98f8b946edd8\") " pod="openstack/kube-state-metrics-0" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.879138 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/62f0c278-8dcb-4e9d-8844-98f8b946edd8-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"62f0c278-8dcb-4e9d-8844-98f8b946edd8\") " pod="openstack/kube-state-metrics-0" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.879217 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6qk52\" (UniqueName: \"kubernetes.io/projected/62f0c278-8dcb-4e9d-8844-98f8b946edd8-kube-api-access-6qk52\") pod \"kube-state-metrics-0\" (UID: \"62f0c278-8dcb-4e9d-8844-98f8b946edd8\") " pod="openstack/kube-state-metrics-0" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.879285 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/62f0c278-8dcb-4e9d-8844-98f8b946edd8-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"62f0c278-8dcb-4e9d-8844-98f8b946edd8\") " pod="openstack/kube-state-metrics-0" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.884741 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/62f0c278-8dcb-4e9d-8844-98f8b946edd8-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"62f0c278-8dcb-4e9d-8844-98f8b946edd8\") " pod="openstack/kube-state-metrics-0" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.887484 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/62f0c278-8dcb-4e9d-8844-98f8b946edd8-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"62f0c278-8dcb-4e9d-8844-98f8b946edd8\") " pod="openstack/kube-state-metrics-0" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.888147 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/62f0c278-8dcb-4e9d-8844-98f8b946edd8-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"62f0c278-8dcb-4e9d-8844-98f8b946edd8\") " pod="openstack/kube-state-metrics-0" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.900360 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6qk52\" (UniqueName: \"kubernetes.io/projected/62f0c278-8dcb-4e9d-8844-98f8b946edd8-kube-api-access-6qk52\") pod \"kube-state-metrics-0\" (UID: \"62f0c278-8dcb-4e9d-8844-98f8b946edd8\") " pod="openstack/kube-state-metrics-0" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.923894 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.992105 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68c677b759-g4lpm"] Dec 09 15:52:02 crc kubenswrapper[4894]: I1209 15:52:02.994706 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-68c677b759-g4lpm" podUID="d3f62001-a311-45d1-ab65-15cd3de84d4e" containerName="dnsmasq-dns" containerID="cri-o://038221b6bd6c60ebdac98f041b56b3384ae28d03a6a88aaee2fad23b522e57b7" gracePeriod=10 Dec 09 15:52:03 crc kubenswrapper[4894]: I1209 15:52:03.135496 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Dec 09 15:52:03 crc kubenswrapper[4894]: I1209 15:52:03.594602 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Dec 09 15:52:03 crc kubenswrapper[4894]: W1209 15:52:03.597101 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod62f0c278_8dcb_4e9d_8844_98f8b946edd8.slice/crio-19871972725a53af99a55b77c322b3689a6045dd36a660ec8402f18e9a162605 WatchSource:0}: Error finding container 19871972725a53af99a55b77c322b3689a6045dd36a660ec8402f18e9a162605: Status 404 returned error can't find the container with id 19871972725a53af99a55b77c322b3689a6045dd36a660ec8402f18e9a162605 Dec 09 15:52:03 crc kubenswrapper[4894]: I1209 15:52:03.615108 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"62f0c278-8dcb-4e9d-8844-98f8b946edd8","Type":"ContainerStarted","Data":"19871972725a53af99a55b77c322b3689a6045dd36a660ec8402f18e9a162605"} Dec 09 15:52:03 crc kubenswrapper[4894]: I1209 15:52:03.616582 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"00ff6ff2-b452-45ae-8f9d-314363bcb126","Type":"ContainerStarted","Data":"f1da57090a0165790834d0b1a3a5fab5c3717555960a6d99189b8c7217b1a0ca"} Dec 09 15:52:03 crc kubenswrapper[4894]: I1209 15:52:03.622754 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb","Type":"ContainerStarted","Data":"62493d8e3b705c3e5b04425414000d458b9741f791db79fa72c7af19236476e3"} Dec 09 15:52:04 crc kubenswrapper[4894]: I1209 15:52:04.119509 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48d401c8-b930-410e-a313-80399a664641" path="/var/lib/kubelet/pods/48d401c8-b930-410e-a313-80399a664641/volumes" Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.656226 4894 generic.go:334] "Generic (PLEG): container finished" podID="7fc523c0-a7fc-4689-a3ed-e43288589249" containerID="7460daec013ac6ac4932f395ede5e82849994c5daabe2e8383db32ee12fc51a9" exitCode=0 Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.656694 4894 generic.go:334] "Generic (PLEG): container finished" podID="7fc523c0-a7fc-4689-a3ed-e43288589249" containerID="860d30ad535d6172272c81311f33b40c1de0b026651c3282663f5518b92bb13a" exitCode=2 Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.656704 4894 generic.go:334] "Generic (PLEG): container finished" podID="7fc523c0-a7fc-4689-a3ed-e43288589249" containerID="67eccfb195452e20fd8a6e68c0898294288d59b21488e44bb1ec66b34cf286f6" exitCode=0 Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.656286 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fc523c0-a7fc-4689-a3ed-e43288589249","Type":"ContainerDied","Data":"7460daec013ac6ac4932f395ede5e82849994c5daabe2e8383db32ee12fc51a9"} Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.656813 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fc523c0-a7fc-4689-a3ed-e43288589249","Type":"ContainerDied","Data":"860d30ad535d6172272c81311f33b40c1de0b026651c3282663f5518b92bb13a"} Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.656829 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fc523c0-a7fc-4689-a3ed-e43288589249","Type":"ContainerDied","Data":"67eccfb195452e20fd8a6e68c0898294288d59b21488e44bb1ec66b34cf286f6"} Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.659086 4894 generic.go:334] "Generic (PLEG): container finished" podID="9418641c-296a-4d30-a02a-52c150f99cd2" containerID="8687a54d3c71b9460c2c6bfe89afe97dd96d5b812daf0a197b0df3633e5b6c66" exitCode=0 Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.659117 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9418641c-296a-4d30-a02a-52c150f99cd2","Type":"ContainerDied","Data":"8687a54d3c71b9460c2c6bfe89afe97dd96d5b812daf0a197b0df3633e5b6c66"} Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.660713 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb","Type":"ContainerStarted","Data":"ef67d9cb479d3583d45a1d6dad7d1d87ad5f2a87d3cc7232eb7e7d8a438d9dbf"} Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.665864 4894 generic.go:334] "Generic (PLEG): container finished" podID="8cefa8f8-0c44-4827-af46-671cea04dd97" containerID="f2556776659526f969cb1e90d6261ee3f6e0696cae9699946f744b0fac60b09b" exitCode=0 Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.666002 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8cefa8f8-0c44-4827-af46-671cea04dd97","Type":"ContainerDied","Data":"f2556776659526f969cb1e90d6261ee3f6e0696cae9699946f744b0fac60b09b"} Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.668413 4894 generic.go:334] "Generic (PLEG): container finished" podID="d3f62001-a311-45d1-ab65-15cd3de84d4e" containerID="038221b6bd6c60ebdac98f041b56b3384ae28d03a6a88aaee2fad23b522e57b7" exitCode=0 Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.668670 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68c677b759-g4lpm" event={"ID":"d3f62001-a311-45d1-ab65-15cd3de84d4e","Type":"ContainerDied","Data":"038221b6bd6c60ebdac98f041b56b3384ae28d03a6a88aaee2fad23b522e57b7"} Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.670752 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"00ff6ff2-b452-45ae-8f9d-314363bcb126","Type":"ContainerStarted","Data":"23fe0e13bd6901fc95fbb3b88533704d056b070fea14d25027e79ba13da7c603"} Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.671152 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.683795 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-68c677b759-g4lpm" podUID="d3f62001-a311-45d1-ab65-15cd3de84d4e" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.156:5353: connect: connection refused" Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.700281 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=4.700252217 podStartE2EDuration="4.700252217s" podCreationTimestamp="2025-12-09 15:52:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:52:05.687566816 +0000 UTC m=+1220.006777485" watchObservedRunningTime="2025-12-09 15:52:05.700252217 +0000 UTC m=+1220.019462886" Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.705353 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.749885 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9418641c-296a-4d30-a02a-52c150f99cd2-combined-ca-bundle\") pod \"9418641c-296a-4d30-a02a-52c150f99cd2\" (UID: \"9418641c-296a-4d30-a02a-52c150f99cd2\") " Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.750030 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swcwt\" (UniqueName: \"kubernetes.io/projected/9418641c-296a-4d30-a02a-52c150f99cd2-kube-api-access-swcwt\") pod \"9418641c-296a-4d30-a02a-52c150f99cd2\" (UID: \"9418641c-296a-4d30-a02a-52c150f99cd2\") " Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.750110 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9418641c-296a-4d30-a02a-52c150f99cd2-logs\") pod \"9418641c-296a-4d30-a02a-52c150f99cd2\" (UID: \"9418641c-296a-4d30-a02a-52c150f99cd2\") " Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.750146 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9418641c-296a-4d30-a02a-52c150f99cd2-config-data\") pod \"9418641c-296a-4d30-a02a-52c150f99cd2\" (UID: \"9418641c-296a-4d30-a02a-52c150f99cd2\") " Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.751484 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9418641c-296a-4d30-a02a-52c150f99cd2-logs" (OuterVolumeSpecName: "logs") pod "9418641c-296a-4d30-a02a-52c150f99cd2" (UID: "9418641c-296a-4d30-a02a-52c150f99cd2"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.761962 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9418641c-296a-4d30-a02a-52c150f99cd2-kube-api-access-swcwt" (OuterVolumeSpecName: "kube-api-access-swcwt") pod "9418641c-296a-4d30-a02a-52c150f99cd2" (UID: "9418641c-296a-4d30-a02a-52c150f99cd2"). InnerVolumeSpecName "kube-api-access-swcwt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.790989 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9418641c-296a-4d30-a02a-52c150f99cd2-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9418641c-296a-4d30-a02a-52c150f99cd2" (UID: "9418641c-296a-4d30-a02a-52c150f99cd2"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.802548 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9418641c-296a-4d30-a02a-52c150f99cd2-config-data" (OuterVolumeSpecName: "config-data") pod "9418641c-296a-4d30-a02a-52c150f99cd2" (UID: "9418641c-296a-4d30-a02a-52c150f99cd2"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.852818 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9418641c-296a-4d30-a02a-52c150f99cd2-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.852967 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swcwt\" (UniqueName: \"kubernetes.io/projected/9418641c-296a-4d30-a02a-52c150f99cd2-kube-api-access-swcwt\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.853078 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9418641c-296a-4d30-a02a-52c150f99cd2-logs\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:05 crc kubenswrapper[4894]: I1209 15:52:05.853270 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9418641c-296a-4d30-a02a-52c150f99cd2-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.154284 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68c677b759-g4lpm" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.258336 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-ovsdbserver-nb\") pod \"d3f62001-a311-45d1-ab65-15cd3de84d4e\" (UID: \"d3f62001-a311-45d1-ab65-15cd3de84d4e\") " Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.258400 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-dns-svc\") pod \"d3f62001-a311-45d1-ab65-15cd3de84d4e\" (UID: \"d3f62001-a311-45d1-ab65-15cd3de84d4e\") " Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.258491 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8h6b7\" (UniqueName: \"kubernetes.io/projected/d3f62001-a311-45d1-ab65-15cd3de84d4e-kube-api-access-8h6b7\") pod \"d3f62001-a311-45d1-ab65-15cd3de84d4e\" (UID: \"d3f62001-a311-45d1-ab65-15cd3de84d4e\") " Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.258534 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-ovsdbserver-sb\") pod \"d3f62001-a311-45d1-ab65-15cd3de84d4e\" (UID: \"d3f62001-a311-45d1-ab65-15cd3de84d4e\") " Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.258552 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-config\") pod \"d3f62001-a311-45d1-ab65-15cd3de84d4e\" (UID: \"d3f62001-a311-45d1-ab65-15cd3de84d4e\") " Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.266836 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d3f62001-a311-45d1-ab65-15cd3de84d4e-kube-api-access-8h6b7" (OuterVolumeSpecName: "kube-api-access-8h6b7") pod "d3f62001-a311-45d1-ab65-15cd3de84d4e" (UID: "d3f62001-a311-45d1-ab65-15cd3de84d4e"). InnerVolumeSpecName "kube-api-access-8h6b7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.310357 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "d3f62001-a311-45d1-ab65-15cd3de84d4e" (UID: "d3f62001-a311-45d1-ab65-15cd3de84d4e"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.311288 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "d3f62001-a311-45d1-ab65-15cd3de84d4e" (UID: "d3f62001-a311-45d1-ab65-15cd3de84d4e"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.315622 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-config" (OuterVolumeSpecName: "config") pod "d3f62001-a311-45d1-ab65-15cd3de84d4e" (UID: "d3f62001-a311-45d1-ab65-15cd3de84d4e"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.315863 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "d3f62001-a311-45d1-ab65-15cd3de84d4e" (UID: "d3f62001-a311-45d1-ab65-15cd3de84d4e"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.360886 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8h6b7\" (UniqueName: \"kubernetes.io/projected/d3f62001-a311-45d1-ab65-15cd3de84d4e-kube-api-access-8h6b7\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.360917 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.360929 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.360938 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.360946 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/d3f62001-a311-45d1-ab65-15cd3de84d4e-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.682925 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9418641c-296a-4d30-a02a-52c150f99cd2","Type":"ContainerDied","Data":"b05d10557cd9a98666a0d567674b620320fbc19d2bcef8c03149816bd99aaa14"} Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.683233 4894 scope.go:117] "RemoveContainer" containerID="8687a54d3c71b9460c2c6bfe89afe97dd96d5b812daf0a197b0df3633e5b6c66" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.682935 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.689989 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb","Type":"ContainerStarted","Data":"8f81a06f5c2089ac5ad882ae5ee1988d11b7bd4006892b97fb4e4d6aaf466596"} Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.694326 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"62f0c278-8dcb-4e9d-8844-98f8b946edd8","Type":"ContainerStarted","Data":"e57ae6e758932742ce162ead924dc6c02f39173d80840c682d35638f86fdc6d6"} Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.694483 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.701709 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-68c677b759-g4lpm" event={"ID":"d3f62001-a311-45d1-ab65-15cd3de84d4e","Type":"ContainerDied","Data":"0f422df104602f440c514ebe0b707da27dafca695e003ca2d519c8d6bbb7e39f"} Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.701883 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-68c677b759-g4lpm" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.710410 4894 scope.go:117] "RemoveContainer" containerID="ec90db2fbe4702580bec2f02502f736637c41093f6c7f277c82047e5b880698f" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.720268 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=5.720247185 podStartE2EDuration="5.720247185s" podCreationTimestamp="2025-12-09 15:52:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:52:06.710460388 +0000 UTC m=+1221.029671077" watchObservedRunningTime="2025-12-09 15:52:06.720247185 +0000 UTC m=+1221.039457854" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.747394 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=3.003366955 podStartE2EDuration="4.747375632s" podCreationTimestamp="2025-12-09 15:52:02 +0000 UTC" firstStartedPulling="2025-12-09 15:52:03.598766506 +0000 UTC m=+1217.917977175" lastFinishedPulling="2025-12-09 15:52:05.342775163 +0000 UTC m=+1219.661985852" observedRunningTime="2025-12-09 15:52:06.73111071 +0000 UTC m=+1221.050321379" watchObservedRunningTime="2025-12-09 15:52:06.747375632 +0000 UTC m=+1221.066586301" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.754824 4894 scope.go:117] "RemoveContainer" containerID="038221b6bd6c60ebdac98f041b56b3384ae28d03a6a88aaee2fad23b522e57b7" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.764937 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.777734 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.787686 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-68c677b759-g4lpm"] Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.799192 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-68c677b759-g4lpm"] Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.824165 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 09 15:52:06 crc kubenswrapper[4894]: E1209 15:52:06.824578 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f62001-a311-45d1-ab65-15cd3de84d4e" containerName="init" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.824592 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f62001-a311-45d1-ab65-15cd3de84d4e" containerName="init" Dec 09 15:52:06 crc kubenswrapper[4894]: E1209 15:52:06.824610 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9418641c-296a-4d30-a02a-52c150f99cd2" containerName="nova-api-api" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.824618 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9418641c-296a-4d30-a02a-52c150f99cd2" containerName="nova-api-api" Dec 09 15:52:06 crc kubenswrapper[4894]: E1209 15:52:06.824657 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9418641c-296a-4d30-a02a-52c150f99cd2" containerName="nova-api-log" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.824663 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9418641c-296a-4d30-a02a-52c150f99cd2" containerName="nova-api-log" Dec 09 15:52:06 crc kubenswrapper[4894]: E1209 15:52:06.824678 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d3f62001-a311-45d1-ab65-15cd3de84d4e" containerName="dnsmasq-dns" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.824685 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="d3f62001-a311-45d1-ab65-15cd3de84d4e" containerName="dnsmasq-dns" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.824875 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9418641c-296a-4d30-a02a-52c150f99cd2" containerName="nova-api-log" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.824896 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="d3f62001-a311-45d1-ab65-15cd3de84d4e" containerName="dnsmasq-dns" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.824905 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9418641c-296a-4d30-a02a-52c150f99cd2" containerName="nova-api-api" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.825793 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.827527 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.838786 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.893815 4894 scope.go:117] "RemoveContainer" containerID="877dd577395d8d411177a9ca04183dcb4a4add06bd2ce7a00dd54e8902046404" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.897599 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.978740 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cefa8f8-0c44-4827-af46-671cea04dd97-combined-ca-bundle\") pod \"8cefa8f8-0c44-4827-af46-671cea04dd97\" (UID: \"8cefa8f8-0c44-4827-af46-671cea04dd97\") " Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.979591 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cefa8f8-0c44-4827-af46-671cea04dd97-config-data\") pod \"8cefa8f8-0c44-4827-af46-671cea04dd97\" (UID: \"8cefa8f8-0c44-4827-af46-671cea04dd97\") " Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.979993 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gv979\" (UniqueName: \"kubernetes.io/projected/8cefa8f8-0c44-4827-af46-671cea04dd97-kube-api-access-gv979\") pod \"8cefa8f8-0c44-4827-af46-671cea04dd97\" (UID: \"8cefa8f8-0c44-4827-af46-671cea04dd97\") " Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.982579 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aef6b29-41e9-4a4d-87a3-d637b71277ec-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7aef6b29-41e9-4a4d-87a3-d637b71277ec\") " pod="openstack/nova-api-0" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.982852 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hghlg\" (UniqueName: \"kubernetes.io/projected/7aef6b29-41e9-4a4d-87a3-d637b71277ec-kube-api-access-hghlg\") pod \"nova-api-0\" (UID: \"7aef6b29-41e9-4a4d-87a3-d637b71277ec\") " pod="openstack/nova-api-0" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.982941 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7aef6b29-41e9-4a4d-87a3-d637b71277ec-logs\") pod \"nova-api-0\" (UID: \"7aef6b29-41e9-4a4d-87a3-d637b71277ec\") " pod="openstack/nova-api-0" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.983154 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aef6b29-41e9-4a4d-87a3-d637b71277ec-config-data\") pod \"nova-api-0\" (UID: \"7aef6b29-41e9-4a4d-87a3-d637b71277ec\") " pod="openstack/nova-api-0" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.983757 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cefa8f8-0c44-4827-af46-671cea04dd97-kube-api-access-gv979" (OuterVolumeSpecName: "kube-api-access-gv979") pod "8cefa8f8-0c44-4827-af46-671cea04dd97" (UID: "8cefa8f8-0c44-4827-af46-671cea04dd97"). InnerVolumeSpecName "kube-api-access-gv979". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:52:06 crc kubenswrapper[4894]: I1209 15:52:06.984744 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gv979\" (UniqueName: \"kubernetes.io/projected/8cefa8f8-0c44-4827-af46-671cea04dd97-kube-api-access-gv979\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.007905 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cefa8f8-0c44-4827-af46-671cea04dd97-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8cefa8f8-0c44-4827-af46-671cea04dd97" (UID: "8cefa8f8-0c44-4827-af46-671cea04dd97"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.010718 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cefa8f8-0c44-4827-af46-671cea04dd97-config-data" (OuterVolumeSpecName: "config-data") pod "8cefa8f8-0c44-4827-af46-671cea04dd97" (UID: "8cefa8f8-0c44-4827-af46-671cea04dd97"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.087215 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hghlg\" (UniqueName: \"kubernetes.io/projected/7aef6b29-41e9-4a4d-87a3-d637b71277ec-kube-api-access-hghlg\") pod \"nova-api-0\" (UID: \"7aef6b29-41e9-4a4d-87a3-d637b71277ec\") " pod="openstack/nova-api-0" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.087281 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7aef6b29-41e9-4a4d-87a3-d637b71277ec-logs\") pod \"nova-api-0\" (UID: \"7aef6b29-41e9-4a4d-87a3-d637b71277ec\") " pod="openstack/nova-api-0" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.087331 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aef6b29-41e9-4a4d-87a3-d637b71277ec-config-data\") pod \"nova-api-0\" (UID: \"7aef6b29-41e9-4a4d-87a3-d637b71277ec\") " pod="openstack/nova-api-0" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.087397 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aef6b29-41e9-4a4d-87a3-d637b71277ec-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7aef6b29-41e9-4a4d-87a3-d637b71277ec\") " pod="openstack/nova-api-0" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.087545 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8cefa8f8-0c44-4827-af46-671cea04dd97-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.087582 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8cefa8f8-0c44-4827-af46-671cea04dd97-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.088787 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7aef6b29-41e9-4a4d-87a3-d637b71277ec-logs\") pod \"nova-api-0\" (UID: \"7aef6b29-41e9-4a4d-87a3-d637b71277ec\") " pod="openstack/nova-api-0" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.092285 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aef6b29-41e9-4a4d-87a3-d637b71277ec-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"7aef6b29-41e9-4a4d-87a3-d637b71277ec\") " pod="openstack/nova-api-0" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.092331 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aef6b29-41e9-4a4d-87a3-d637b71277ec-config-data\") pod \"nova-api-0\" (UID: \"7aef6b29-41e9-4a4d-87a3-d637b71277ec\") " pod="openstack/nova-api-0" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.110261 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hghlg\" (UniqueName: \"kubernetes.io/projected/7aef6b29-41e9-4a4d-87a3-d637b71277ec-kube-api-access-hghlg\") pod \"nova-api-0\" (UID: \"7aef6b29-41e9-4a4d-87a3-d637b71277ec\") " pod="openstack/nova-api-0" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.202122 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.287154 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.287371 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.669294 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.730127 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7aef6b29-41e9-4a4d-87a3-d637b71277ec","Type":"ContainerStarted","Data":"e588ffbbac93af97c985e29dd009d1e996229368eb5025c2709f518272c4cdc5"} Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.734059 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"8cefa8f8-0c44-4827-af46-671cea04dd97","Type":"ContainerDied","Data":"e45489f2964965311cd4e86cc33d4f71da390cc4c562f46b582e2cd534322a83"} Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.734116 4894 scope.go:117] "RemoveContainer" containerID="f2556776659526f969cb1e90d6261ee3f6e0696cae9699946f744b0fac60b09b" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.734331 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.744040 4894 generic.go:334] "Generic (PLEG): container finished" podID="7fc523c0-a7fc-4689-a3ed-e43288589249" containerID="df527646f6cc5b4c20571dd4d6a938bc0029167739028aac8ab7f819a664e895" exitCode=0 Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.744915 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fc523c0-a7fc-4689-a3ed-e43288589249","Type":"ContainerDied","Data":"df527646f6cc5b4c20571dd4d6a938bc0029167739028aac8ab7f819a664e895"} Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.909348 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.948172 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.967821 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.989894 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 15:52:07 crc kubenswrapper[4894]: E1209 15:52:07.990344 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc523c0-a7fc-4689-a3ed-e43288589249" containerName="sg-core" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.990366 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc523c0-a7fc-4689-a3ed-e43288589249" containerName="sg-core" Dec 09 15:52:07 crc kubenswrapper[4894]: E1209 15:52:07.990386 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc523c0-a7fc-4689-a3ed-e43288589249" containerName="proxy-httpd" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.990394 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc523c0-a7fc-4689-a3ed-e43288589249" containerName="proxy-httpd" Dec 09 15:52:07 crc kubenswrapper[4894]: E1209 15:52:07.990412 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8cefa8f8-0c44-4827-af46-671cea04dd97" containerName="nova-scheduler-scheduler" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.990421 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8cefa8f8-0c44-4827-af46-671cea04dd97" containerName="nova-scheduler-scheduler" Dec 09 15:52:07 crc kubenswrapper[4894]: E1209 15:52:07.990435 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc523c0-a7fc-4689-a3ed-e43288589249" containerName="ceilometer-notification-agent" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.990443 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc523c0-a7fc-4689-a3ed-e43288589249" containerName="ceilometer-notification-agent" Dec 09 15:52:07 crc kubenswrapper[4894]: E1209 15:52:07.990463 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fc523c0-a7fc-4689-a3ed-e43288589249" containerName="ceilometer-central-agent" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.990470 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fc523c0-a7fc-4689-a3ed-e43288589249" containerName="ceilometer-central-agent" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.990682 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fc523c0-a7fc-4689-a3ed-e43288589249" containerName="ceilometer-central-agent" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.990703 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fc523c0-a7fc-4689-a3ed-e43288589249" containerName="sg-core" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.990728 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fc523c0-a7fc-4689-a3ed-e43288589249" containerName="proxy-httpd" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.990749 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fc523c0-a7fc-4689-a3ed-e43288589249" containerName="ceilometer-notification-agent" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.990763 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8cefa8f8-0c44-4827-af46-671cea04dd97" containerName="nova-scheduler-scheduler" Dec 09 15:52:07 crc kubenswrapper[4894]: I1209 15:52:07.991482 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.025681 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.030200 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-combined-ca-bundle\") pod \"7fc523c0-a7fc-4689-a3ed-e43288589249\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.030265 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-config-data\") pod \"7fc523c0-a7fc-4689-a3ed-e43288589249\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.030311 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fc523c0-a7fc-4689-a3ed-e43288589249-log-httpd\") pod \"7fc523c0-a7fc-4689-a3ed-e43288589249\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.030412 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-sg-core-conf-yaml\") pod \"7fc523c0-a7fc-4689-a3ed-e43288589249\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.030430 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-275rd\" (UniqueName: \"kubernetes.io/projected/7fc523c0-a7fc-4689-a3ed-e43288589249-kube-api-access-275rd\") pod \"7fc523c0-a7fc-4689-a3ed-e43288589249\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.030477 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-scripts\") pod \"7fc523c0-a7fc-4689-a3ed-e43288589249\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.030494 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fc523c0-a7fc-4689-a3ed-e43288589249-run-httpd\") pod \"7fc523c0-a7fc-4689-a3ed-e43288589249\" (UID: \"7fc523c0-a7fc-4689-a3ed-e43288589249\") " Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.032061 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fc523c0-a7fc-4689-a3ed-e43288589249-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "7fc523c0-a7fc-4689-a3ed-e43288589249" (UID: "7fc523c0-a7fc-4689-a3ed-e43288589249"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.033068 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.036045 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7fc523c0-a7fc-4689-a3ed-e43288589249-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "7fc523c0-a7fc-4689-a3ed-e43288589249" (UID: "7fc523c0-a7fc-4689-a3ed-e43288589249"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.051932 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fc523c0-a7fc-4689-a3ed-e43288589249-kube-api-access-275rd" (OuterVolumeSpecName: "kube-api-access-275rd") pod "7fc523c0-a7fc-4689-a3ed-e43288589249" (UID: "7fc523c0-a7fc-4689-a3ed-e43288589249"). InnerVolumeSpecName "kube-api-access-275rd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.074842 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-scripts" (OuterVolumeSpecName: "scripts") pod "7fc523c0-a7fc-4689-a3ed-e43288589249" (UID: "7fc523c0-a7fc-4689-a3ed-e43288589249"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.131756 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a185555-06a5-4630-8d9c-700c44af003a-config-data\") pod \"nova-scheduler-0\" (UID: \"6a185555-06a5-4630-8d9c-700c44af003a\") " pod="openstack/nova-scheduler-0" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.132077 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-74p29\" (UniqueName: \"kubernetes.io/projected/6a185555-06a5-4630-8d9c-700c44af003a-kube-api-access-74p29\") pod \"nova-scheduler-0\" (UID: \"6a185555-06a5-4630-8d9c-700c44af003a\") " pod="openstack/nova-scheduler-0" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.132200 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a185555-06a5-4630-8d9c-700c44af003a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6a185555-06a5-4630-8d9c-700c44af003a\") " pod="openstack/nova-scheduler-0" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.132339 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-275rd\" (UniqueName: \"kubernetes.io/projected/7fc523c0-a7fc-4689-a3ed-e43288589249-kube-api-access-275rd\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.132417 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.132500 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fc523c0-a7fc-4689-a3ed-e43288589249-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.132577 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/7fc523c0-a7fc-4689-a3ed-e43288589249-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.151853 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cefa8f8-0c44-4827-af46-671cea04dd97" path="/var/lib/kubelet/pods/8cefa8f8-0c44-4827-af46-671cea04dd97/volumes" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.155588 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9418641c-296a-4d30-a02a-52c150f99cd2" path="/var/lib/kubelet/pods/9418641c-296a-4d30-a02a-52c150f99cd2/volumes" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.156571 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d3f62001-a311-45d1-ab65-15cd3de84d4e" path="/var/lib/kubelet/pods/d3f62001-a311-45d1-ab65-15cd3de84d4e/volumes" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.175901 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "7fc523c0-a7fc-4689-a3ed-e43288589249" (UID: "7fc523c0-a7fc-4689-a3ed-e43288589249"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.229965 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7fc523c0-a7fc-4689-a3ed-e43288589249" (UID: "7fc523c0-a7fc-4689-a3ed-e43288589249"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.234411 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a185555-06a5-4630-8d9c-700c44af003a-config-data\") pod \"nova-scheduler-0\" (UID: \"6a185555-06a5-4630-8d9c-700c44af003a\") " pod="openstack/nova-scheduler-0" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.234794 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-74p29\" (UniqueName: \"kubernetes.io/projected/6a185555-06a5-4630-8d9c-700c44af003a-kube-api-access-74p29\") pod \"nova-scheduler-0\" (UID: \"6a185555-06a5-4630-8d9c-700c44af003a\") " pod="openstack/nova-scheduler-0" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.235414 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a185555-06a5-4630-8d9c-700c44af003a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6a185555-06a5-4630-8d9c-700c44af003a\") " pod="openstack/nova-scheduler-0" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.235603 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.235622 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.238112 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a185555-06a5-4630-8d9c-700c44af003a-config-data\") pod \"nova-scheduler-0\" (UID: \"6a185555-06a5-4630-8d9c-700c44af003a\") " pod="openstack/nova-scheduler-0" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.239092 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a185555-06a5-4630-8d9c-700c44af003a-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"6a185555-06a5-4630-8d9c-700c44af003a\") " pod="openstack/nova-scheduler-0" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.247571 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-config-data" (OuterVolumeSpecName: "config-data") pod "7fc523c0-a7fc-4689-a3ed-e43288589249" (UID: "7fc523c0-a7fc-4689-a3ed-e43288589249"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.252076 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-74p29\" (UniqueName: \"kubernetes.io/projected/6a185555-06a5-4630-8d9c-700c44af003a-kube-api-access-74p29\") pod \"nova-scheduler-0\" (UID: \"6a185555-06a5-4630-8d9c-700c44af003a\") " pod="openstack/nova-scheduler-0" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.334898 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.348779 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fc523c0-a7fc-4689-a3ed-e43288589249-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.757542 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.757761 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"7fc523c0-a7fc-4689-a3ed-e43288589249","Type":"ContainerDied","Data":"864a53b4fb77e885f8b056a0f8962402af3af1c26c8f441bc88311d8776769f7"} Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.758303 4894 scope.go:117] "RemoveContainer" containerID="7460daec013ac6ac4932f395ede5e82849994c5daabe2e8383db32ee12fc51a9" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.768613 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7aef6b29-41e9-4a4d-87a3-d637b71277ec","Type":"ContainerStarted","Data":"edf29a77f679a3cd876792adbe1937377c0d30749e1d475ce665092ddc89a7a7"} Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.787812 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 15:52:08 crc kubenswrapper[4894]: W1209 15:52:08.796829 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6a185555_06a5_4630_8d9c_700c44af003a.slice/crio-767349dea3cea5a4aef57e7cfcc98adcfb0a2efb9c62e85af517260abde611cb WatchSource:0}: Error finding container 767349dea3cea5a4aef57e7cfcc98adcfb0a2efb9c62e85af517260abde611cb: Status 404 returned error can't find the container with id 767349dea3cea5a4aef57e7cfcc98adcfb0a2efb9c62e85af517260abde611cb Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.815470 4894 scope.go:117] "RemoveContainer" containerID="860d30ad535d6172272c81311f33b40c1de0b026651c3282663f5518b92bb13a" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.848586 4894 scope.go:117] "RemoveContainer" containerID="df527646f6cc5b4c20571dd4d6a938bc0029167739028aac8ab7f819a664e895" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.853842 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.873277 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.882935 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.885872 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.891267 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.891481 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.891831 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.893270 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:52:08 crc kubenswrapper[4894]: I1209 15:52:08.917907 4894 scope.go:117] "RemoveContainer" containerID="67eccfb195452e20fd8a6e68c0898294288d59b21488e44bb1ec66b34cf286f6" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.064568 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b78a7533-956b-43e9-8652-476a8bc5b223-run-httpd\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.064621 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.064782 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-scripts\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.064848 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.064878 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcprx\" (UniqueName: \"kubernetes.io/projected/b78a7533-956b-43e9-8652-476a8bc5b223-kube-api-access-gcprx\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.064900 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-config-data\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.064915 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.064930 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b78a7533-956b-43e9-8652-476a8bc5b223-log-httpd\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.167259 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-scripts\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.167346 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.167376 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcprx\" (UniqueName: \"kubernetes.io/projected/b78a7533-956b-43e9-8652-476a8bc5b223-kube-api-access-gcprx\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.167401 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-config-data\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.167420 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.167446 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b78a7533-956b-43e9-8652-476a8bc5b223-log-httpd\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.167473 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b78a7533-956b-43e9-8652-476a8bc5b223-run-httpd\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.167506 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.168091 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b78a7533-956b-43e9-8652-476a8bc5b223-run-httpd\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.168648 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b78a7533-956b-43e9-8652-476a8bc5b223-log-httpd\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.177213 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.180574 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-scripts\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.181728 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.181959 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-config-data\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.184609 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.190511 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcprx\" (UniqueName: \"kubernetes.io/projected/b78a7533-956b-43e9-8652-476a8bc5b223-kube-api-access-gcprx\") pod \"ceilometer-0\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.207188 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.671515 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.782664 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6a185555-06a5-4630-8d9c-700c44af003a","Type":"ContainerStarted","Data":"d50a6675c092a631fc70b8d36a691aea8241c786f666ebeacf9ea9b4d56c78db"} Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.783815 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6a185555-06a5-4630-8d9c-700c44af003a","Type":"ContainerStarted","Data":"767349dea3cea5a4aef57e7cfcc98adcfb0a2efb9c62e85af517260abde611cb"} Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.787301 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7aef6b29-41e9-4a4d-87a3-d637b71277ec","Type":"ContainerStarted","Data":"02ea8037c9d9ceb7de41b82fa4c7665ca83107471e82e01f5564d0a53a394e93"} Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.788957 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b78a7533-956b-43e9-8652-476a8bc5b223","Type":"ContainerStarted","Data":"a784f6d92f1534c6084f8ed79cc4539cbb898493038d30aa51ca7622c7f3db1c"} Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.807136 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.807113459 podStartE2EDuration="2.807113459s" podCreationTimestamp="2025-12-09 15:52:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:52:09.797899926 +0000 UTC m=+1224.117110615" watchObservedRunningTime="2025-12-09 15:52:09.807113459 +0000 UTC m=+1224.126324128" Dec 09 15:52:09 crc kubenswrapper[4894]: I1209 15:52:09.835152 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.835132577 podStartE2EDuration="3.835132577s" podCreationTimestamp="2025-12-09 15:52:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:52:09.827333331 +0000 UTC m=+1224.146544000" watchObservedRunningTime="2025-12-09 15:52:09.835132577 +0000 UTC m=+1224.154343246" Dec 09 15:52:10 crc kubenswrapper[4894]: I1209 15:52:10.116765 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fc523c0-a7fc-4689-a3ed-e43288589249" path="/var/lib/kubelet/pods/7fc523c0-a7fc-4689-a3ed-e43288589249/volumes" Dec 09 15:52:10 crc kubenswrapper[4894]: I1209 15:52:10.802881 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b78a7533-956b-43e9-8652-476a8bc5b223","Type":"ContainerStarted","Data":"a3beab5fa382653ddaedc45562fbb28480cac908a344bc88e52048f2e5271c1c"} Dec 09 15:52:11 crc kubenswrapper[4894]: I1209 15:52:11.729984 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 15:52:11 crc kubenswrapper[4894]: I1209 15:52:11.813065 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b78a7533-956b-43e9-8652-476a8bc5b223","Type":"ContainerStarted","Data":"33e0170ef5eb4ddf49671db46b48095a9e0c11a1657e251a494c63e815cf1c56"} Dec 09 15:52:11 crc kubenswrapper[4894]: I1209 15:52:11.813107 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b78a7533-956b-43e9-8652-476a8bc5b223","Type":"ContainerStarted","Data":"08eee7ee8e4e9a10046703b22147205680f3ad6058ae7ef5273567476b3977ca"} Dec 09 15:52:12 crc kubenswrapper[4894]: I1209 15:52:12.189965 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Dec 09 15:52:12 crc kubenswrapper[4894]: I1209 15:52:12.286797 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 09 15:52:12 crc kubenswrapper[4894]: I1209 15:52:12.286850 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 09 15:52:13 crc kubenswrapper[4894]: I1209 15:52:13.148405 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Dec 09 15:52:13 crc kubenswrapper[4894]: I1209 15:52:13.299769 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e97b390d-e40e-45d2-8eb7-c4ff7205e5fb" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.184:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 15:52:13 crc kubenswrapper[4894]: I1209 15:52:13.300026 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="e97b390d-e40e-45d2-8eb7-c4ff7205e5fb" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.184:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 15:52:13 crc kubenswrapper[4894]: I1209 15:52:13.335365 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 09 15:52:14 crc kubenswrapper[4894]: I1209 15:52:14.841282 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b78a7533-956b-43e9-8652-476a8bc5b223","Type":"ContainerStarted","Data":"45a5e8c6763d5dbab6a5f75acf00f62250fed7761b42ec7057cd272340485991"} Dec 09 15:52:14 crc kubenswrapper[4894]: I1209 15:52:14.842836 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 15:52:14 crc kubenswrapper[4894]: I1209 15:52:14.868760 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.397891237 podStartE2EDuration="6.868738208s" podCreationTimestamp="2025-12-09 15:52:08 +0000 UTC" firstStartedPulling="2025-12-09 15:52:09.675753575 +0000 UTC m=+1223.994964244" lastFinishedPulling="2025-12-09 15:52:14.146600546 +0000 UTC m=+1228.465811215" observedRunningTime="2025-12-09 15:52:14.862453279 +0000 UTC m=+1229.181663948" watchObservedRunningTime="2025-12-09 15:52:14.868738208 +0000 UTC m=+1229.187948877" Dec 09 15:52:17 crc kubenswrapper[4894]: I1209 15:52:17.203338 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 15:52:17 crc kubenswrapper[4894]: I1209 15:52:17.203604 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 15:52:18 crc kubenswrapper[4894]: I1209 15:52:18.288956 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7aef6b29-41e9-4a4d-87a3-d637b71277ec" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 15:52:18 crc kubenswrapper[4894]: I1209 15:52:18.290489 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="7aef6b29-41e9-4a4d-87a3-d637b71277ec" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.186:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Dec 09 15:52:18 crc kubenswrapper[4894]: I1209 15:52:18.336042 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 09 15:52:18 crc kubenswrapper[4894]: I1209 15:52:18.364318 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 09 15:52:18 crc kubenswrapper[4894]: I1209 15:52:18.900892 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 09 15:52:22 crc kubenswrapper[4894]: I1209 15:52:22.292000 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 09 15:52:22 crc kubenswrapper[4894]: I1209 15:52:22.294259 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 09 15:52:22 crc kubenswrapper[4894]: I1209 15:52:22.298000 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 09 15:52:22 crc kubenswrapper[4894]: I1209 15:52:22.921115 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 09 15:52:24 crc kubenswrapper[4894]: I1209 15:52:24.860501 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:24 crc kubenswrapper[4894]: I1209 15:52:24.931322 4894 generic.go:334] "Generic (PLEG): container finished" podID="ee76fef6-16b9-49e7-adb6-2145cc0e5e03" containerID="2633f428a1d9bbb3c94cbe02ce3dc5ad819ef17b00d7689b8ead36a81c92ec74" exitCode=137 Dec 09 15:52:24 crc kubenswrapper[4894]: I1209 15:52:24.931390 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:24 crc kubenswrapper[4894]: I1209 15:52:24.931443 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ee76fef6-16b9-49e7-adb6-2145cc0e5e03","Type":"ContainerDied","Data":"2633f428a1d9bbb3c94cbe02ce3dc5ad819ef17b00d7689b8ead36a81c92ec74"} Dec 09 15:52:24 crc kubenswrapper[4894]: I1209 15:52:24.931488 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"ee76fef6-16b9-49e7-adb6-2145cc0e5e03","Type":"ContainerDied","Data":"17f38d262e9e0e9200c407f0f8c07a7276297df704f7c6565e8574e5f1d33a6f"} Dec 09 15:52:24 crc kubenswrapper[4894]: I1209 15:52:24.931508 4894 scope.go:117] "RemoveContainer" containerID="2633f428a1d9bbb3c94cbe02ce3dc5ad819ef17b00d7689b8ead36a81c92ec74" Dec 09 15:52:24 crc kubenswrapper[4894]: I1209 15:52:24.962127 4894 scope.go:117] "RemoveContainer" containerID="2633f428a1d9bbb3c94cbe02ce3dc5ad819ef17b00d7689b8ead36a81c92ec74" Dec 09 15:52:24 crc kubenswrapper[4894]: E1209 15:52:24.962509 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2633f428a1d9bbb3c94cbe02ce3dc5ad819ef17b00d7689b8ead36a81c92ec74\": container with ID starting with 2633f428a1d9bbb3c94cbe02ce3dc5ad819ef17b00d7689b8ead36a81c92ec74 not found: ID does not exist" containerID="2633f428a1d9bbb3c94cbe02ce3dc5ad819ef17b00d7689b8ead36a81c92ec74" Dec 09 15:52:24 crc kubenswrapper[4894]: I1209 15:52:24.962537 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2633f428a1d9bbb3c94cbe02ce3dc5ad819ef17b00d7689b8ead36a81c92ec74"} err="failed to get container status \"2633f428a1d9bbb3c94cbe02ce3dc5ad819ef17b00d7689b8ead36a81c92ec74\": rpc error: code = NotFound desc = could not find container \"2633f428a1d9bbb3c94cbe02ce3dc5ad819ef17b00d7689b8ead36a81c92ec74\": container with ID starting with 2633f428a1d9bbb3c94cbe02ce3dc5ad819ef17b00d7689b8ead36a81c92ec74 not found: ID does not exist" Dec 09 15:52:24 crc kubenswrapper[4894]: I1209 15:52:24.968611 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee76fef6-16b9-49e7-adb6-2145cc0e5e03-config-data\") pod \"ee76fef6-16b9-49e7-adb6-2145cc0e5e03\" (UID: \"ee76fef6-16b9-49e7-adb6-2145cc0e5e03\") " Dec 09 15:52:24 crc kubenswrapper[4894]: I1209 15:52:24.968681 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee76fef6-16b9-49e7-adb6-2145cc0e5e03-combined-ca-bundle\") pod \"ee76fef6-16b9-49e7-adb6-2145cc0e5e03\" (UID: \"ee76fef6-16b9-49e7-adb6-2145cc0e5e03\") " Dec 09 15:52:24 crc kubenswrapper[4894]: I1209 15:52:24.968846 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wjjf\" (UniqueName: \"kubernetes.io/projected/ee76fef6-16b9-49e7-adb6-2145cc0e5e03-kube-api-access-6wjjf\") pod \"ee76fef6-16b9-49e7-adb6-2145cc0e5e03\" (UID: \"ee76fef6-16b9-49e7-adb6-2145cc0e5e03\") " Dec 09 15:52:24 crc kubenswrapper[4894]: I1209 15:52:24.973569 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ee76fef6-16b9-49e7-adb6-2145cc0e5e03-kube-api-access-6wjjf" (OuterVolumeSpecName: "kube-api-access-6wjjf") pod "ee76fef6-16b9-49e7-adb6-2145cc0e5e03" (UID: "ee76fef6-16b9-49e7-adb6-2145cc0e5e03"). InnerVolumeSpecName "kube-api-access-6wjjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:52:24 crc kubenswrapper[4894]: I1209 15:52:24.997861 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee76fef6-16b9-49e7-adb6-2145cc0e5e03-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ee76fef6-16b9-49e7-adb6-2145cc0e5e03" (UID: "ee76fef6-16b9-49e7-adb6-2145cc0e5e03"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.001988 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ee76fef6-16b9-49e7-adb6-2145cc0e5e03-config-data" (OuterVolumeSpecName: "config-data") pod "ee76fef6-16b9-49e7-adb6-2145cc0e5e03" (UID: "ee76fef6-16b9-49e7-adb6-2145cc0e5e03"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.070958 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ee76fef6-16b9-49e7-adb6-2145cc0e5e03-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.071002 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ee76fef6-16b9-49e7-adb6-2145cc0e5e03-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.071060 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wjjf\" (UniqueName: \"kubernetes.io/projected/ee76fef6-16b9-49e7-adb6-2145cc0e5e03-kube-api-access-6wjjf\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.277369 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.295045 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.303729 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 15:52:25 crc kubenswrapper[4894]: E1209 15:52:25.304230 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ee76fef6-16b9-49e7-adb6-2145cc0e5e03" containerName="nova-cell1-novncproxy-novncproxy" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.304255 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ee76fef6-16b9-49e7-adb6-2145cc0e5e03" containerName="nova-cell1-novncproxy-novncproxy" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.304503 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="ee76fef6-16b9-49e7-adb6-2145cc0e5e03" containerName="nova-cell1-novncproxy-novncproxy" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.305302 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.307397 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.311003 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.312217 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.314605 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.376360 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6219c187-146d-45f2-bc64-95639530b601-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6219c187-146d-45f2-bc64-95639530b601\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.376408 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6219c187-146d-45f2-bc64-95639530b601-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6219c187-146d-45f2-bc64-95639530b601\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.376455 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6219c187-146d-45f2-bc64-95639530b601-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6219c187-146d-45f2-bc64-95639530b601\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.376803 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mxhfw\" (UniqueName: \"kubernetes.io/projected/6219c187-146d-45f2-bc64-95639530b601-kube-api-access-mxhfw\") pod \"nova-cell1-novncproxy-0\" (UID: \"6219c187-146d-45f2-bc64-95639530b601\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.377135 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6219c187-146d-45f2-bc64-95639530b601-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6219c187-146d-45f2-bc64-95639530b601\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.479009 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6219c187-146d-45f2-bc64-95639530b601-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6219c187-146d-45f2-bc64-95639530b601\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.479121 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6219c187-146d-45f2-bc64-95639530b601-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6219c187-146d-45f2-bc64-95639530b601\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.479214 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6219c187-146d-45f2-bc64-95639530b601-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6219c187-146d-45f2-bc64-95639530b601\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.479313 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6219c187-146d-45f2-bc64-95639530b601-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6219c187-146d-45f2-bc64-95639530b601\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.479483 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mxhfw\" (UniqueName: \"kubernetes.io/projected/6219c187-146d-45f2-bc64-95639530b601-kube-api-access-mxhfw\") pod \"nova-cell1-novncproxy-0\" (UID: \"6219c187-146d-45f2-bc64-95639530b601\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.483777 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6219c187-146d-45f2-bc64-95639530b601-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"6219c187-146d-45f2-bc64-95639530b601\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.484615 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/6219c187-146d-45f2-bc64-95639530b601-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6219c187-146d-45f2-bc64-95639530b601\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.485006 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/6219c187-146d-45f2-bc64-95639530b601-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"6219c187-146d-45f2-bc64-95639530b601\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.485027 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6219c187-146d-45f2-bc64-95639530b601-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"6219c187-146d-45f2-bc64-95639530b601\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.495442 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mxhfw\" (UniqueName: \"kubernetes.io/projected/6219c187-146d-45f2-bc64-95639530b601-kube-api-access-mxhfw\") pod \"nova-cell1-novncproxy-0\" (UID: \"6219c187-146d-45f2-bc64-95639530b601\") " pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:25 crc kubenswrapper[4894]: I1209 15:52:25.623916 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:26 crc kubenswrapper[4894]: I1209 15:52:26.063685 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Dec 09 15:52:26 crc kubenswrapper[4894]: I1209 15:52:26.135163 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ee76fef6-16b9-49e7-adb6-2145cc0e5e03" path="/var/lib/kubelet/pods/ee76fef6-16b9-49e7-adb6-2145cc0e5e03/volumes" Dec 09 15:52:26 crc kubenswrapper[4894]: I1209 15:52:26.953836 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6219c187-146d-45f2-bc64-95639530b601","Type":"ContainerStarted","Data":"9b29252568442c7dcb9326046ce8acc659ce336e3ba34d6aad36e2e87727579d"} Dec 09 15:52:26 crc kubenswrapper[4894]: I1209 15:52:26.954317 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"6219c187-146d-45f2-bc64-95639530b601","Type":"ContainerStarted","Data":"6879a6f821e9a2f44bd1519963daa95065dc2311c6df1a4136d2dcb303695626"} Dec 09 15:52:26 crc kubenswrapper[4894]: I1209 15:52:26.972376 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=1.972337895 podStartE2EDuration="1.972337895s" podCreationTimestamp="2025-12-09 15:52:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:52:26.971115832 +0000 UTC m=+1241.290326501" watchObservedRunningTime="2025-12-09 15:52:26.972337895 +0000 UTC m=+1241.291548564" Dec 09 15:52:27 crc kubenswrapper[4894]: I1209 15:52:27.206932 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 09 15:52:27 crc kubenswrapper[4894]: I1209 15:52:27.207549 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 09 15:52:27 crc kubenswrapper[4894]: I1209 15:52:27.211092 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 09 15:52:27 crc kubenswrapper[4894]: I1209 15:52:27.211155 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 09 15:52:27 crc kubenswrapper[4894]: I1209 15:52:27.960162 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 09 15:52:27 crc kubenswrapper[4894]: I1209 15:52:27.963229 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 09 15:52:28 crc kubenswrapper[4894]: I1209 15:52:28.160732 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6f69c5c76f-kcfmr"] Dec 09 15:52:28 crc kubenswrapper[4894]: I1209 15:52:28.162436 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" Dec 09 15:52:28 crc kubenswrapper[4894]: I1209 15:52:28.190355 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f69c5c76f-kcfmr"] Dec 09 15:52:28 crc kubenswrapper[4894]: I1209 15:52:28.226146 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-ovsdbserver-sb\") pod \"dnsmasq-dns-6f69c5c76f-kcfmr\" (UID: \"70af2821-3434-450b-b5e1-d876b1ec90ed\") " pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" Dec 09 15:52:28 crc kubenswrapper[4894]: I1209 15:52:28.226221 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-ovsdbserver-nb\") pod \"dnsmasq-dns-6f69c5c76f-kcfmr\" (UID: \"70af2821-3434-450b-b5e1-d876b1ec90ed\") " pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" Dec 09 15:52:28 crc kubenswrapper[4894]: I1209 15:52:28.226286 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j2rvm\" (UniqueName: \"kubernetes.io/projected/70af2821-3434-450b-b5e1-d876b1ec90ed-kube-api-access-j2rvm\") pod \"dnsmasq-dns-6f69c5c76f-kcfmr\" (UID: \"70af2821-3434-450b-b5e1-d876b1ec90ed\") " pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" Dec 09 15:52:28 crc kubenswrapper[4894]: I1209 15:52:28.226316 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-config\") pod \"dnsmasq-dns-6f69c5c76f-kcfmr\" (UID: \"70af2821-3434-450b-b5e1-d876b1ec90ed\") " pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" Dec 09 15:52:28 crc kubenswrapper[4894]: I1209 15:52:28.226387 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-dns-svc\") pod \"dnsmasq-dns-6f69c5c76f-kcfmr\" (UID: \"70af2821-3434-450b-b5e1-d876b1ec90ed\") " pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" Dec 09 15:52:28 crc kubenswrapper[4894]: I1209 15:52:28.327989 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j2rvm\" (UniqueName: \"kubernetes.io/projected/70af2821-3434-450b-b5e1-d876b1ec90ed-kube-api-access-j2rvm\") pod \"dnsmasq-dns-6f69c5c76f-kcfmr\" (UID: \"70af2821-3434-450b-b5e1-d876b1ec90ed\") " pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" Dec 09 15:52:28 crc kubenswrapper[4894]: I1209 15:52:28.328166 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-config\") pod \"dnsmasq-dns-6f69c5c76f-kcfmr\" (UID: \"70af2821-3434-450b-b5e1-d876b1ec90ed\") " pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" Dec 09 15:52:28 crc kubenswrapper[4894]: I1209 15:52:28.328268 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-dns-svc\") pod \"dnsmasq-dns-6f69c5c76f-kcfmr\" (UID: \"70af2821-3434-450b-b5e1-d876b1ec90ed\") " pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" Dec 09 15:52:28 crc kubenswrapper[4894]: I1209 15:52:28.328360 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-ovsdbserver-sb\") pod \"dnsmasq-dns-6f69c5c76f-kcfmr\" (UID: \"70af2821-3434-450b-b5e1-d876b1ec90ed\") " pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" Dec 09 15:52:28 crc kubenswrapper[4894]: I1209 15:52:28.328402 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-ovsdbserver-nb\") pod \"dnsmasq-dns-6f69c5c76f-kcfmr\" (UID: \"70af2821-3434-450b-b5e1-d876b1ec90ed\") " pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" Dec 09 15:52:28 crc kubenswrapper[4894]: I1209 15:52:28.329219 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-config\") pod \"dnsmasq-dns-6f69c5c76f-kcfmr\" (UID: \"70af2821-3434-450b-b5e1-d876b1ec90ed\") " pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" Dec 09 15:52:28 crc kubenswrapper[4894]: I1209 15:52:28.329390 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-ovsdbserver-nb\") pod \"dnsmasq-dns-6f69c5c76f-kcfmr\" (UID: \"70af2821-3434-450b-b5e1-d876b1ec90ed\") " pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" Dec 09 15:52:28 crc kubenswrapper[4894]: I1209 15:52:28.329394 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-ovsdbserver-sb\") pod \"dnsmasq-dns-6f69c5c76f-kcfmr\" (UID: \"70af2821-3434-450b-b5e1-d876b1ec90ed\") " pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" Dec 09 15:52:28 crc kubenswrapper[4894]: I1209 15:52:28.329424 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-dns-svc\") pod \"dnsmasq-dns-6f69c5c76f-kcfmr\" (UID: \"70af2821-3434-450b-b5e1-d876b1ec90ed\") " pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" Dec 09 15:52:28 crc kubenswrapper[4894]: I1209 15:52:28.349607 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j2rvm\" (UniqueName: \"kubernetes.io/projected/70af2821-3434-450b-b5e1-d876b1ec90ed-kube-api-access-j2rvm\") pod \"dnsmasq-dns-6f69c5c76f-kcfmr\" (UID: \"70af2821-3434-450b-b5e1-d876b1ec90ed\") " pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" Dec 09 15:52:28 crc kubenswrapper[4894]: I1209 15:52:28.481029 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" Dec 09 15:52:28 crc kubenswrapper[4894]: W1209 15:52:28.968903 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod70af2821_3434_450b_b5e1_d876b1ec90ed.slice/crio-84b274a4c36e1454582a33d579d1e29d04f3278ec53a4df6dfc782a13d54a1bb WatchSource:0}: Error finding container 84b274a4c36e1454582a33d579d1e29d04f3278ec53a4df6dfc782a13d54a1bb: Status 404 returned error can't find the container with id 84b274a4c36e1454582a33d579d1e29d04f3278ec53a4df6dfc782a13d54a1bb Dec 09 15:52:28 crc kubenswrapper[4894]: I1209 15:52:28.983584 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6f69c5c76f-kcfmr"] Dec 09 15:52:29 crc kubenswrapper[4894]: I1209 15:52:29.989709 4894 generic.go:334] "Generic (PLEG): container finished" podID="70af2821-3434-450b-b5e1-d876b1ec90ed" containerID="5e78c3b981df33475f764aa60405f26aa363f841a5465e8fd5ff95a290ef4fa9" exitCode=0 Dec 09 15:52:29 crc kubenswrapper[4894]: I1209 15:52:29.989777 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" event={"ID":"70af2821-3434-450b-b5e1-d876b1ec90ed","Type":"ContainerDied","Data":"5e78c3b981df33475f764aa60405f26aa363f841a5465e8fd5ff95a290ef4fa9"} Dec 09 15:52:29 crc kubenswrapper[4894]: I1209 15:52:29.990076 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" event={"ID":"70af2821-3434-450b-b5e1-d876b1ec90ed","Type":"ContainerStarted","Data":"84b274a4c36e1454582a33d579d1e29d04f3278ec53a4df6dfc782a13d54a1bb"} Dec 09 15:52:30 crc kubenswrapper[4894]: I1209 15:52:30.255578 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:52:30 crc kubenswrapper[4894]: I1209 15:52:30.256271 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b78a7533-956b-43e9-8652-476a8bc5b223" containerName="ceilometer-central-agent" containerID="cri-o://a3beab5fa382653ddaedc45562fbb28480cac908a344bc88e52048f2e5271c1c" gracePeriod=30 Dec 09 15:52:30 crc kubenswrapper[4894]: I1209 15:52:30.256369 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b78a7533-956b-43e9-8652-476a8bc5b223" containerName="sg-core" containerID="cri-o://33e0170ef5eb4ddf49671db46b48095a9e0c11a1657e251a494c63e815cf1c56" gracePeriod=30 Dec 09 15:52:30 crc kubenswrapper[4894]: I1209 15:52:30.256429 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b78a7533-956b-43e9-8652-476a8bc5b223" containerName="ceilometer-notification-agent" containerID="cri-o://08eee7ee8e4e9a10046703b22147205680f3ad6058ae7ef5273567476b3977ca" gracePeriod=30 Dec 09 15:52:30 crc kubenswrapper[4894]: I1209 15:52:30.256598 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="b78a7533-956b-43e9-8652-476a8bc5b223" containerName="proxy-httpd" containerID="cri-o://45a5e8c6763d5dbab6a5f75acf00f62250fed7761b42ec7057cd272340485991" gracePeriod=30 Dec 09 15:52:30 crc kubenswrapper[4894]: I1209 15:52:30.272991 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="b78a7533-956b-43e9-8652-476a8bc5b223" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 502" Dec 09 15:52:30 crc kubenswrapper[4894]: I1209 15:52:30.624697 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:30 crc kubenswrapper[4894]: I1209 15:52:30.933792 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 15:52:30 crc kubenswrapper[4894]: I1209 15:52:30.997447 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7aef6b29-41e9-4a4d-87a3-d637b71277ec" containerName="nova-api-log" containerID="cri-o://edf29a77f679a3cd876792adbe1937377c0d30749e1d475ce665092ddc89a7a7" gracePeriod=30 Dec 09 15:52:30 crc kubenswrapper[4894]: I1209 15:52:30.997608 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="7aef6b29-41e9-4a4d-87a3-d637b71277ec" containerName="nova-api-api" containerID="cri-o://02ea8037c9d9ceb7de41b82fa4c7665ca83107471e82e01f5564d0a53a394e93" gracePeriod=30 Dec 09 15:52:34 crc kubenswrapper[4894]: I1209 15:52:34.040580 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b78a7533-956b-43e9-8652-476a8bc5b223","Type":"ContainerDied","Data":"33e0170ef5eb4ddf49671db46b48095a9e0c11a1657e251a494c63e815cf1c56"} Dec 09 15:52:34 crc kubenswrapper[4894]: I1209 15:52:34.040552 4894 generic.go:334] "Generic (PLEG): container finished" podID="b78a7533-956b-43e9-8652-476a8bc5b223" containerID="33e0170ef5eb4ddf49671db46b48095a9e0c11a1657e251a494c63e815cf1c56" exitCode=2 Dec 09 15:52:35 crc kubenswrapper[4894]: I1209 15:52:35.054232 4894 generic.go:334] "Generic (PLEG): container finished" podID="b78a7533-956b-43e9-8652-476a8bc5b223" containerID="45a5e8c6763d5dbab6a5f75acf00f62250fed7761b42ec7057cd272340485991" exitCode=0 Dec 09 15:52:35 crc kubenswrapper[4894]: I1209 15:52:35.054549 4894 generic.go:334] "Generic (PLEG): container finished" podID="b78a7533-956b-43e9-8652-476a8bc5b223" containerID="08eee7ee8e4e9a10046703b22147205680f3ad6058ae7ef5273567476b3977ca" exitCode=0 Dec 09 15:52:35 crc kubenswrapper[4894]: I1209 15:52:35.054558 4894 generic.go:334] "Generic (PLEG): container finished" podID="b78a7533-956b-43e9-8652-476a8bc5b223" containerID="a3beab5fa382653ddaedc45562fbb28480cac908a344bc88e52048f2e5271c1c" exitCode=0 Dec 09 15:52:35 crc kubenswrapper[4894]: I1209 15:52:35.054307 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b78a7533-956b-43e9-8652-476a8bc5b223","Type":"ContainerDied","Data":"45a5e8c6763d5dbab6a5f75acf00f62250fed7761b42ec7057cd272340485991"} Dec 09 15:52:35 crc kubenswrapper[4894]: I1209 15:52:35.054618 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b78a7533-956b-43e9-8652-476a8bc5b223","Type":"ContainerDied","Data":"08eee7ee8e4e9a10046703b22147205680f3ad6058ae7ef5273567476b3977ca"} Dec 09 15:52:35 crc kubenswrapper[4894]: I1209 15:52:35.054631 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b78a7533-956b-43e9-8652-476a8bc5b223","Type":"ContainerDied","Data":"a3beab5fa382653ddaedc45562fbb28480cac908a344bc88e52048f2e5271c1c"} Dec 09 15:52:35 crc kubenswrapper[4894]: I1209 15:52:35.057802 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" event={"ID":"70af2821-3434-450b-b5e1-d876b1ec90ed","Type":"ContainerStarted","Data":"aa6796049964581c30dd9bf1a036c7e6210221c7c2c7b9c76fdcdcbec36a2055"} Dec 09 15:52:35 crc kubenswrapper[4894]: I1209 15:52:35.057940 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" Dec 09 15:52:35 crc kubenswrapper[4894]: I1209 15:52:35.061099 4894 generic.go:334] "Generic (PLEG): container finished" podID="7aef6b29-41e9-4a4d-87a3-d637b71277ec" containerID="02ea8037c9d9ceb7de41b82fa4c7665ca83107471e82e01f5564d0a53a394e93" exitCode=0 Dec 09 15:52:35 crc kubenswrapper[4894]: I1209 15:52:35.061117 4894 generic.go:334] "Generic (PLEG): container finished" podID="7aef6b29-41e9-4a4d-87a3-d637b71277ec" containerID="edf29a77f679a3cd876792adbe1937377c0d30749e1d475ce665092ddc89a7a7" exitCode=143 Dec 09 15:52:35 crc kubenswrapper[4894]: I1209 15:52:35.061145 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7aef6b29-41e9-4a4d-87a3-d637b71277ec","Type":"ContainerDied","Data":"02ea8037c9d9ceb7de41b82fa4c7665ca83107471e82e01f5564d0a53a394e93"} Dec 09 15:52:35 crc kubenswrapper[4894]: I1209 15:52:35.061198 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7aef6b29-41e9-4a4d-87a3-d637b71277ec","Type":"ContainerDied","Data":"edf29a77f679a3cd876792adbe1937377c0d30749e1d475ce665092ddc89a7a7"} Dec 09 15:52:35 crc kubenswrapper[4894]: I1209 15:52:35.086436 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" podStartSLOduration=7.086416371 podStartE2EDuration="7.086416371s" podCreationTimestamp="2025-12-09 15:52:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:52:35.085330281 +0000 UTC m=+1249.404540970" watchObservedRunningTime="2025-12-09 15:52:35.086416371 +0000 UTC m=+1249.405627080" Dec 09 15:52:35 crc kubenswrapper[4894]: I1209 15:52:35.624623 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:35 crc kubenswrapper[4894]: I1209 15:52:35.643420 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.008606 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.020863 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.071059 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.071072 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"7aef6b29-41e9-4a4d-87a3-d637b71277ec","Type":"ContainerDied","Data":"e588ffbbac93af97c985e29dd009d1e996229368eb5025c2709f518272c4cdc5"} Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.071166 4894 scope.go:117] "RemoveContainer" containerID="02ea8037c9d9ceb7de41b82fa4c7665ca83107471e82e01f5564d0a53a394e93" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.076623 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7aef6b29-41e9-4a4d-87a3-d637b71277ec-logs\") pod \"7aef6b29-41e9-4a4d-87a3-d637b71277ec\" (UID: \"7aef6b29-41e9-4a4d-87a3-d637b71277ec\") " Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.076767 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hghlg\" (UniqueName: \"kubernetes.io/projected/7aef6b29-41e9-4a4d-87a3-d637b71277ec-kube-api-access-hghlg\") pod \"7aef6b29-41e9-4a4d-87a3-d637b71277ec\" (UID: \"7aef6b29-41e9-4a4d-87a3-d637b71277ec\") " Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.076909 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aef6b29-41e9-4a4d-87a3-d637b71277ec-config-data\") pod \"7aef6b29-41e9-4a4d-87a3-d637b71277ec\" (UID: \"7aef6b29-41e9-4a4d-87a3-d637b71277ec\") " Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.077026 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aef6b29-41e9-4a4d-87a3-d637b71277ec-combined-ca-bundle\") pod \"7aef6b29-41e9-4a4d-87a3-d637b71277ec\" (UID: \"7aef6b29-41e9-4a4d-87a3-d637b71277ec\") " Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.077342 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7aef6b29-41e9-4a4d-87a3-d637b71277ec-logs" (OuterVolumeSpecName: "logs") pod "7aef6b29-41e9-4a4d-87a3-d637b71277ec" (UID: "7aef6b29-41e9-4a4d-87a3-d637b71277ec"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.077671 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7aef6b29-41e9-4a4d-87a3-d637b71277ec-logs\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.084793 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7aef6b29-41e9-4a4d-87a3-d637b71277ec-kube-api-access-hghlg" (OuterVolumeSpecName: "kube-api-access-hghlg") pod "7aef6b29-41e9-4a4d-87a3-d637b71277ec" (UID: "7aef6b29-41e9-4a4d-87a3-d637b71277ec"). InnerVolumeSpecName "kube-api-access-hghlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.095190 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.096094 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"b78a7533-956b-43e9-8652-476a8bc5b223","Type":"ContainerDied","Data":"a784f6d92f1534c6084f8ed79cc4539cbb898493038d30aa51ca7622c7f3db1c"} Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.135058 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aef6b29-41e9-4a4d-87a3-d637b71277ec-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7aef6b29-41e9-4a4d-87a3-d637b71277ec" (UID: "7aef6b29-41e9-4a4d-87a3-d637b71277ec"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.137690 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7aef6b29-41e9-4a4d-87a3-d637b71277ec-config-data" (OuterVolumeSpecName: "config-data") pod "7aef6b29-41e9-4a4d-87a3-d637b71277ec" (UID: "7aef6b29-41e9-4a4d-87a3-d637b71277ec"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.178358 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-config-data\") pod \"b78a7533-956b-43e9-8652-476a8bc5b223\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.178424 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-combined-ca-bundle\") pod \"b78a7533-956b-43e9-8652-476a8bc5b223\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.178474 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b78a7533-956b-43e9-8652-476a8bc5b223-log-httpd\") pod \"b78a7533-956b-43e9-8652-476a8bc5b223\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.178525 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-ceilometer-tls-certs\") pod \"b78a7533-956b-43e9-8652-476a8bc5b223\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.178578 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-scripts\") pod \"b78a7533-956b-43e9-8652-476a8bc5b223\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.178609 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcprx\" (UniqueName: \"kubernetes.io/projected/b78a7533-956b-43e9-8652-476a8bc5b223-kube-api-access-gcprx\") pod \"b78a7533-956b-43e9-8652-476a8bc5b223\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.178661 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b78a7533-956b-43e9-8652-476a8bc5b223-run-httpd\") pod \"b78a7533-956b-43e9-8652-476a8bc5b223\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.178752 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-sg-core-conf-yaml\") pod \"b78a7533-956b-43e9-8652-476a8bc5b223\" (UID: \"b78a7533-956b-43e9-8652-476a8bc5b223\") " Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.179211 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7aef6b29-41e9-4a4d-87a3-d637b71277ec-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.179225 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hghlg\" (UniqueName: \"kubernetes.io/projected/7aef6b29-41e9-4a4d-87a3-d637b71277ec-kube-api-access-hghlg\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.179238 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7aef6b29-41e9-4a4d-87a3-d637b71277ec-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.179433 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b78a7533-956b-43e9-8652-476a8bc5b223-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "b78a7533-956b-43e9-8652-476a8bc5b223" (UID: "b78a7533-956b-43e9-8652-476a8bc5b223"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.182587 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b78a7533-956b-43e9-8652-476a8bc5b223-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "b78a7533-956b-43e9-8652-476a8bc5b223" (UID: "b78a7533-956b-43e9-8652-476a8bc5b223"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.184672 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-scripts" (OuterVolumeSpecName: "scripts") pod "b78a7533-956b-43e9-8652-476a8bc5b223" (UID: "b78a7533-956b-43e9-8652-476a8bc5b223"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.187277 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b78a7533-956b-43e9-8652-476a8bc5b223-kube-api-access-gcprx" (OuterVolumeSpecName: "kube-api-access-gcprx") pod "b78a7533-956b-43e9-8652-476a8bc5b223" (UID: "b78a7533-956b-43e9-8652-476a8bc5b223"). InnerVolumeSpecName "kube-api-access-gcprx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.189261 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.221612 4894 scope.go:117] "RemoveContainer" containerID="edf29a77f679a3cd876792adbe1937377c0d30749e1d475ce665092ddc89a7a7" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.224468 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "b78a7533-956b-43e9-8652-476a8bc5b223" (UID: "b78a7533-956b-43e9-8652-476a8bc5b223"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.264732 4894 scope.go:117] "RemoveContainer" containerID="45a5e8c6763d5dbab6a5f75acf00f62250fed7761b42ec7057cd272340485991" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.283999 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcprx\" (UniqueName: \"kubernetes.io/projected/b78a7533-956b-43e9-8652-476a8bc5b223-kube-api-access-gcprx\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.284038 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b78a7533-956b-43e9-8652-476a8bc5b223-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.284052 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.284065 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/b78a7533-956b-43e9-8652-476a8bc5b223-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.284076 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.285053 4894 scope.go:117] "RemoveContainer" containerID="33e0170ef5eb4ddf49671db46b48095a9e0c11a1657e251a494c63e815cf1c56" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.287228 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b78a7533-956b-43e9-8652-476a8bc5b223" (UID: "b78a7533-956b-43e9-8652-476a8bc5b223"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.300012 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "b78a7533-956b-43e9-8652-476a8bc5b223" (UID: "b78a7533-956b-43e9-8652-476a8bc5b223"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.304626 4894 scope.go:117] "RemoveContainer" containerID="08eee7ee8e4e9a10046703b22147205680f3ad6058ae7ef5273567476b3977ca" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.344034 4894 scope.go:117] "RemoveContainer" containerID="a3beab5fa382653ddaedc45562fbb28480cac908a344bc88e52048f2e5271c1c" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.351665 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-kwtnn"] Dec 09 15:52:36 crc kubenswrapper[4894]: E1209 15:52:36.352130 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7aef6b29-41e9-4a4d-87a3-d637b71277ec" containerName="nova-api-api" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.352157 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7aef6b29-41e9-4a4d-87a3-d637b71277ec" containerName="nova-api-api" Dec 09 15:52:36 crc kubenswrapper[4894]: E1209 15:52:36.352185 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b78a7533-956b-43e9-8652-476a8bc5b223" containerName="ceilometer-central-agent" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.352194 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b78a7533-956b-43e9-8652-476a8bc5b223" containerName="ceilometer-central-agent" Dec 09 15:52:36 crc kubenswrapper[4894]: E1209 15:52:36.352215 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b78a7533-956b-43e9-8652-476a8bc5b223" containerName="sg-core" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.352224 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b78a7533-956b-43e9-8652-476a8bc5b223" containerName="sg-core" Dec 09 15:52:36 crc kubenswrapper[4894]: E1209 15:52:36.352243 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7aef6b29-41e9-4a4d-87a3-d637b71277ec" containerName="nova-api-log" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.352251 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7aef6b29-41e9-4a4d-87a3-d637b71277ec" containerName="nova-api-log" Dec 09 15:52:36 crc kubenswrapper[4894]: E1209 15:52:36.352262 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b78a7533-956b-43e9-8652-476a8bc5b223" containerName="proxy-httpd" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.352269 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b78a7533-956b-43e9-8652-476a8bc5b223" containerName="proxy-httpd" Dec 09 15:52:36 crc kubenswrapper[4894]: E1209 15:52:36.352282 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b78a7533-956b-43e9-8652-476a8bc5b223" containerName="ceilometer-notification-agent" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.352290 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b78a7533-956b-43e9-8652-476a8bc5b223" containerName="ceilometer-notification-agent" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.352511 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7aef6b29-41e9-4a4d-87a3-d637b71277ec" containerName="nova-api-api" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.352532 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b78a7533-956b-43e9-8652-476a8bc5b223" containerName="ceilometer-central-agent" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.352545 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b78a7533-956b-43e9-8652-476a8bc5b223" containerName="ceilometer-notification-agent" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.352561 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b78a7533-956b-43e9-8652-476a8bc5b223" containerName="sg-core" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.352573 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b78a7533-956b-43e9-8652-476a8bc5b223" containerName="proxy-httpd" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.352589 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7aef6b29-41e9-4a4d-87a3-d637b71277ec" containerName="nova-api-log" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.353362 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kwtnn" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.355902 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.356736 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.370319 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-config-data" (OuterVolumeSpecName: "config-data") pod "b78a7533-956b-43e9-8652-476a8bc5b223" (UID: "b78a7533-956b-43e9-8652-476a8bc5b223"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.380947 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-kwtnn"] Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.390232 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.391837 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.391973 4894 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/b78a7533-956b-43e9-8652-476a8bc5b223-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.443349 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.460699 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.469804 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.471607 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.478072 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.478979 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.479134 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.481106 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.490073 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.493026 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95bff0ec-6b68-439a-b478-de6549452662-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-kwtnn\" (UID: \"95bff0ec-6b68-439a-b478-de6549452662\") " pod="openstack/nova-cell1-cell-mapping-kwtnn" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.493091 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skbx7\" (UniqueName: \"kubernetes.io/projected/95bff0ec-6b68-439a-b478-de6549452662-kube-api-access-skbx7\") pod \"nova-cell1-cell-mapping-kwtnn\" (UID: \"95bff0ec-6b68-439a-b478-de6549452662\") " pod="openstack/nova-cell1-cell-mapping-kwtnn" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.493131 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95bff0ec-6b68-439a-b478-de6549452662-config-data\") pod \"nova-cell1-cell-mapping-kwtnn\" (UID: \"95bff0ec-6b68-439a-b478-de6549452662\") " pod="openstack/nova-cell1-cell-mapping-kwtnn" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.493177 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95bff0ec-6b68-439a-b478-de6549452662-scripts\") pod \"nova-cell1-cell-mapping-kwtnn\" (UID: \"95bff0ec-6b68-439a-b478-de6549452662\") " pod="openstack/nova-cell1-cell-mapping-kwtnn" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.509933 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.517902 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.520238 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.522984 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.523148 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.523309 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.539790 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.598729 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e6f0669-9599-480e-aa0a-9b36bfd851ff-logs\") pod \"nova-api-0\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " pod="openstack/nova-api-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.598923 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-public-tls-certs\") pod \"nova-api-0\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " pod="openstack/nova-api-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.598995 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-run-httpd\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.599057 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.599095 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xxlzn\" (UniqueName: \"kubernetes.io/projected/9e6f0669-9599-480e-aa0a-9b36bfd851ff-kube-api-access-xxlzn\") pod \"nova-api-0\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " pod="openstack/nova-api-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.599127 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95bff0ec-6b68-439a-b478-de6549452662-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-kwtnn\" (UID: \"95bff0ec-6b68-439a-b478-de6549452662\") " pod="openstack/nova-cell1-cell-mapping-kwtnn" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.599254 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skbx7\" (UniqueName: \"kubernetes.io/projected/95bff0ec-6b68-439a-b478-de6549452662-kube-api-access-skbx7\") pod \"nova-cell1-cell-mapping-kwtnn\" (UID: \"95bff0ec-6b68-439a-b478-de6549452662\") " pod="openstack/nova-cell1-cell-mapping-kwtnn" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.599276 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-scripts\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.599324 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95bff0ec-6b68-439a-b478-de6549452662-config-data\") pod \"nova-cell1-cell-mapping-kwtnn\" (UID: \"95bff0ec-6b68-439a-b478-de6549452662\") " pod="openstack/nova-cell1-cell-mapping-kwtnn" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.599417 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95bff0ec-6b68-439a-b478-de6549452662-scripts\") pod \"nova-cell1-cell-mapping-kwtnn\" (UID: \"95bff0ec-6b68-439a-b478-de6549452662\") " pod="openstack/nova-cell1-cell-mapping-kwtnn" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.599458 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " pod="openstack/nova-api-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.599507 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-config-data\") pod \"nova-api-0\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " pod="openstack/nova-api-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.599533 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.599563 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " pod="openstack/nova-api-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.599592 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-log-httpd\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.599615 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-config-data\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.599671 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkvdh\" (UniqueName: \"kubernetes.io/projected/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-kube-api-access-jkvdh\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.599726 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.603306 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95bff0ec-6b68-439a-b478-de6549452662-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-kwtnn\" (UID: \"95bff0ec-6b68-439a-b478-de6549452662\") " pod="openstack/nova-cell1-cell-mapping-kwtnn" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.603878 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95bff0ec-6b68-439a-b478-de6549452662-config-data\") pod \"nova-cell1-cell-mapping-kwtnn\" (UID: \"95bff0ec-6b68-439a-b478-de6549452662\") " pod="openstack/nova-cell1-cell-mapping-kwtnn" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.604308 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95bff0ec-6b68-439a-b478-de6549452662-scripts\") pod \"nova-cell1-cell-mapping-kwtnn\" (UID: \"95bff0ec-6b68-439a-b478-de6549452662\") " pod="openstack/nova-cell1-cell-mapping-kwtnn" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.618072 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skbx7\" (UniqueName: \"kubernetes.io/projected/95bff0ec-6b68-439a-b478-de6549452662-kube-api-access-skbx7\") pod \"nova-cell1-cell-mapping-kwtnn\" (UID: \"95bff0ec-6b68-439a-b478-de6549452662\") " pod="openstack/nova-cell1-cell-mapping-kwtnn" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.679330 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kwtnn" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.701300 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e6f0669-9599-480e-aa0a-9b36bfd851ff-logs\") pod \"nova-api-0\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " pod="openstack/nova-api-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.701400 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-public-tls-certs\") pod \"nova-api-0\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " pod="openstack/nova-api-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.701436 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-run-httpd\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.701464 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.701487 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xxlzn\" (UniqueName: \"kubernetes.io/projected/9e6f0669-9599-480e-aa0a-9b36bfd851ff-kube-api-access-xxlzn\") pod \"nova-api-0\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " pod="openstack/nova-api-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.701571 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-scripts\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.701629 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " pod="openstack/nova-api-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.701676 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-config-data\") pod \"nova-api-0\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " pod="openstack/nova-api-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.701700 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.701723 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " pod="openstack/nova-api-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.701751 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-log-httpd\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.701773 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-config-data\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.701796 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkvdh\" (UniqueName: \"kubernetes.io/projected/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-kube-api-access-jkvdh\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.701825 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.701994 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-run-httpd\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.702287 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e6f0669-9599-480e-aa0a-9b36bfd851ff-logs\") pod \"nova-api-0\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " pod="openstack/nova-api-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.704383 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-log-httpd\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.705437 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.707475 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-config-data\") pod \"nova-api-0\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " pod="openstack/nova-api-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.708457 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-config-data\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.708625 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.709125 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-public-tls-certs\") pod \"nova-api-0\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " pod="openstack/nova-api-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.710369 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-scripts\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.710816 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.712070 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " pod="openstack/nova-api-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.715356 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-internal-tls-certs\") pod \"nova-api-0\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " pod="openstack/nova-api-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.719854 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xxlzn\" (UniqueName: \"kubernetes.io/projected/9e6f0669-9599-480e-aa0a-9b36bfd851ff-kube-api-access-xxlzn\") pod \"nova-api-0\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " pod="openstack/nova-api-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.722325 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkvdh\" (UniqueName: \"kubernetes.io/projected/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-kube-api-access-jkvdh\") pod \"ceilometer-0\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " pod="openstack/ceilometer-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.796569 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.822229 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:52:36 crc kubenswrapper[4894]: I1209 15:52:36.823234 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:52:37 crc kubenswrapper[4894]: I1209 15:52:37.183306 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-kwtnn"] Dec 09 15:52:37 crc kubenswrapper[4894]: I1209 15:52:37.195197 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:52:37 crc kubenswrapper[4894]: I1209 15:52:37.429986 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 15:52:38 crc kubenswrapper[4894]: I1209 15:52:38.117303 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7aef6b29-41e9-4a4d-87a3-d637b71277ec" path="/var/lib/kubelet/pods/7aef6b29-41e9-4a4d-87a3-d637b71277ec/volumes" Dec 09 15:52:38 crc kubenswrapper[4894]: I1209 15:52:38.118500 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b78a7533-956b-43e9-8652-476a8bc5b223" path="/var/lib/kubelet/pods/b78a7533-956b-43e9-8652-476a8bc5b223/volumes" Dec 09 15:52:38 crc kubenswrapper[4894]: I1209 15:52:38.143862 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kwtnn" event={"ID":"95bff0ec-6b68-439a-b478-de6549452662","Type":"ContainerStarted","Data":"05878ba3c9a9a398a2afffd857ceedb589bc02e084d330c155e3c537a8230aa9"} Dec 09 15:52:38 crc kubenswrapper[4894]: I1209 15:52:38.143913 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kwtnn" event={"ID":"95bff0ec-6b68-439a-b478-de6549452662","Type":"ContainerStarted","Data":"6e0b5acc8eec627950bfe2318efb9eeecfcc8f1aec14137bf77e265fab953d81"} Dec 09 15:52:38 crc kubenswrapper[4894]: I1209 15:52:38.145225 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53","Type":"ContainerStarted","Data":"ba0da5669d8e28ffee7fdce63c6659cfbe1333ce1e4db695fbb8969c536acb47"} Dec 09 15:52:38 crc kubenswrapper[4894]: I1209 15:52:38.147245 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9e6f0669-9599-480e-aa0a-9b36bfd851ff","Type":"ContainerStarted","Data":"e6cd81a4a234df461b1ea3e319fdfcf6d16fa0a898e255df8d62a7ba0847b35a"} Dec 09 15:52:38 crc kubenswrapper[4894]: I1209 15:52:38.147280 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9e6f0669-9599-480e-aa0a-9b36bfd851ff","Type":"ContainerStarted","Data":"4fe547028bba0b3a911c6a4a336af0666449ceed03554255958fc031849e6587"} Dec 09 15:52:38 crc kubenswrapper[4894]: I1209 15:52:38.147291 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9e6f0669-9599-480e-aa0a-9b36bfd851ff","Type":"ContainerStarted","Data":"37e638f3418ec964ac15866c8143137f24cc30c47f41979ac5530f0e43642bf5"} Dec 09 15:52:38 crc kubenswrapper[4894]: I1209 15:52:38.174837 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-kwtnn" podStartSLOduration=2.174818307 podStartE2EDuration="2.174818307s" podCreationTimestamp="2025-12-09 15:52:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:52:38.158094256 +0000 UTC m=+1252.477304925" watchObservedRunningTime="2025-12-09 15:52:38.174818307 +0000 UTC m=+1252.494028976" Dec 09 15:52:38 crc kubenswrapper[4894]: I1209 15:52:38.187698 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.187674854 podStartE2EDuration="2.187674854s" podCreationTimestamp="2025-12-09 15:52:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:52:38.175379502 +0000 UTC m=+1252.494590171" watchObservedRunningTime="2025-12-09 15:52:38.187674854 +0000 UTC m=+1252.506885533" Dec 09 15:52:40 crc kubenswrapper[4894]: I1209 15:52:40.171225 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53","Type":"ContainerStarted","Data":"ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194"} Dec 09 15:52:41 crc kubenswrapper[4894]: I1209 15:52:41.189522 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53","Type":"ContainerStarted","Data":"b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838"} Dec 09 15:52:42 crc kubenswrapper[4894]: I1209 15:52:42.209307 4894 generic.go:334] "Generic (PLEG): container finished" podID="95bff0ec-6b68-439a-b478-de6549452662" containerID="05878ba3c9a9a398a2afffd857ceedb589bc02e084d330c155e3c537a8230aa9" exitCode=0 Dec 09 15:52:42 crc kubenswrapper[4894]: I1209 15:52:42.209381 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kwtnn" event={"ID":"95bff0ec-6b68-439a-b478-de6549452662","Type":"ContainerDied","Data":"05878ba3c9a9a398a2afffd857ceedb589bc02e084d330c155e3c537a8230aa9"} Dec 09 15:52:42 crc kubenswrapper[4894]: I1209 15:52:42.211904 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53","Type":"ContainerStarted","Data":"a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f"} Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.235193 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" containerName="ceilometer-central-agent" containerID="cri-o://ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194" gracePeriod=30 Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.235615 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53","Type":"ContainerStarted","Data":"7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810"} Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.235667 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.235950 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" containerName="proxy-httpd" containerID="cri-o://7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810" gracePeriod=30 Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.235991 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" containerName="sg-core" containerID="cri-o://a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f" gracePeriod=30 Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.236025 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" containerName="ceilometer-notification-agent" containerID="cri-o://b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838" gracePeriod=30 Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.273778 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.641441599 podStartE2EDuration="7.273756344s" podCreationTimestamp="2025-12-09 15:52:36 +0000 UTC" firstStartedPulling="2025-12-09 15:52:37.196358565 +0000 UTC m=+1251.515569234" lastFinishedPulling="2025-12-09 15:52:42.82867331 +0000 UTC m=+1257.147883979" observedRunningTime="2025-12-09 15:52:43.264090824 +0000 UTC m=+1257.583301483" watchObservedRunningTime="2025-12-09 15:52:43.273756344 +0000 UTC m=+1257.592967013" Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.485824 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.553873 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8d97cbc7-jdzrx"] Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.554204 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" podUID="f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5" containerName="dnsmasq-dns" containerID="cri-o://49883e5ea70af46c31fe682fee954dba8d75f41207f6cc81ea8f694214b908c9" gracePeriod=10 Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.709611 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kwtnn" Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.855248 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skbx7\" (UniqueName: \"kubernetes.io/projected/95bff0ec-6b68-439a-b478-de6549452662-kube-api-access-skbx7\") pod \"95bff0ec-6b68-439a-b478-de6549452662\" (UID: \"95bff0ec-6b68-439a-b478-de6549452662\") " Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.855338 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95bff0ec-6b68-439a-b478-de6549452662-config-data\") pod \"95bff0ec-6b68-439a-b478-de6549452662\" (UID: \"95bff0ec-6b68-439a-b478-de6549452662\") " Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.855448 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95bff0ec-6b68-439a-b478-de6549452662-scripts\") pod \"95bff0ec-6b68-439a-b478-de6549452662\" (UID: \"95bff0ec-6b68-439a-b478-de6549452662\") " Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.855490 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95bff0ec-6b68-439a-b478-de6549452662-combined-ca-bundle\") pod \"95bff0ec-6b68-439a-b478-de6549452662\" (UID: \"95bff0ec-6b68-439a-b478-de6549452662\") " Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.862298 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95bff0ec-6b68-439a-b478-de6549452662-scripts" (OuterVolumeSpecName: "scripts") pod "95bff0ec-6b68-439a-b478-de6549452662" (UID: "95bff0ec-6b68-439a-b478-de6549452662"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.869044 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95bff0ec-6b68-439a-b478-de6549452662-kube-api-access-skbx7" (OuterVolumeSpecName: "kube-api-access-skbx7") pod "95bff0ec-6b68-439a-b478-de6549452662" (UID: "95bff0ec-6b68-439a-b478-de6549452662"). InnerVolumeSpecName "kube-api-access-skbx7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.907387 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95bff0ec-6b68-439a-b478-de6549452662-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95bff0ec-6b68-439a-b478-de6549452662" (UID: "95bff0ec-6b68-439a-b478-de6549452662"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.911834 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95bff0ec-6b68-439a-b478-de6549452662-config-data" (OuterVolumeSpecName: "config-data") pod "95bff0ec-6b68-439a-b478-de6549452662" (UID: "95bff0ec-6b68-439a-b478-de6549452662"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.969766 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95bff0ec-6b68-439a-b478-de6549452662-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.969796 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95bff0ec-6b68-439a-b478-de6549452662-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.969808 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skbx7\" (UniqueName: \"kubernetes.io/projected/95bff0ec-6b68-439a-b478-de6549452662-kube-api-access-skbx7\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:43 crc kubenswrapper[4894]: I1209 15:52:43.969817 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/95bff0ec-6b68-439a-b478-de6549452662-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.093585 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.099985 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.171818 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-combined-ca-bundle\") pod \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.171879 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkvdh\" (UniqueName: \"kubernetes.io/projected/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-kube-api-access-jkvdh\") pod \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.171945 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-scripts\") pod \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.171988 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-ceilometer-tls-certs\") pod \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.172051 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-config-data\") pod \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.172068 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-run-httpd\") pod \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.172111 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-sg-core-conf-yaml\") pod \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.172150 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-log-httpd\") pod \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\" (UID: \"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53\") " Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.173174 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" (UID: "41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.173450 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" (UID: "41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.195443 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-kube-api-access-jkvdh" (OuterVolumeSpecName: "kube-api-access-jkvdh") pod "41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" (UID: "41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53"). InnerVolumeSpecName "kube-api-access-jkvdh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.204836 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-scripts" (OuterVolumeSpecName: "scripts") pod "41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" (UID: "41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.217810 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" (UID: "41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.252175 4894 generic.go:334] "Generic (PLEG): container finished" podID="f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5" containerID="49883e5ea70af46c31fe682fee954dba8d75f41207f6cc81ea8f694214b908c9" exitCode=0 Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.252271 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" event={"ID":"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5","Type":"ContainerDied","Data":"49883e5ea70af46c31fe682fee954dba8d75f41207f6cc81ea8f694214b908c9"} Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.252352 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" event={"ID":"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5","Type":"ContainerDied","Data":"c4a267a612dcb97055ff6cf0c4470d0463a41d78a809b8cb0956117b58c226c0"} Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.252373 4894 scope.go:117] "RemoveContainer" containerID="49883e5ea70af46c31fe682fee954dba8d75f41207f6cc81ea8f694214b908c9" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.252603 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8d97cbc7-jdzrx" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.256819 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-kwtnn" event={"ID":"95bff0ec-6b68-439a-b478-de6549452662","Type":"ContainerDied","Data":"6e0b5acc8eec627950bfe2318efb9eeecfcc8f1aec14137bf77e265fab953d81"} Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.256873 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6e0b5acc8eec627950bfe2318efb9eeecfcc8f1aec14137bf77e265fab953d81" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.257043 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-kwtnn" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.261009 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" (UID: "41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.262006 4894 generic.go:334] "Generic (PLEG): container finished" podID="41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" containerID="7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810" exitCode=0 Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.262068 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.262043 4894 generic.go:334] "Generic (PLEG): container finished" podID="41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" containerID="a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f" exitCode=2 Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.262088 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53","Type":"ContainerDied","Data":"7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810"} Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.262124 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53","Type":"ContainerDied","Data":"a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f"} Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.262138 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53","Type":"ContainerDied","Data":"b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838"} Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.262102 4894 generic.go:334] "Generic (PLEG): container finished" podID="41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" containerID="b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838" exitCode=0 Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.262159 4894 generic.go:334] "Generic (PLEG): container finished" podID="41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" containerID="ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194" exitCode=0 Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.262178 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53","Type":"ContainerDied","Data":"ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194"} Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.262186 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53","Type":"ContainerDied","Data":"ba0da5669d8e28ffee7fdce63c6659cfbe1333ce1e4db695fbb8969c536acb47"} Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.266146 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" (UID: "41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.274339 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-ovsdbserver-nb\") pod \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\" (UID: \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\") " Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.274401 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gl4wb\" (UniqueName: \"kubernetes.io/projected/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-kube-api-access-gl4wb\") pod \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\" (UID: \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\") " Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.274466 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-config\") pod \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\" (UID: \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\") " Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.274614 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-dns-svc\") pod \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\" (UID: \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\") " Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.275191 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-ovsdbserver-sb\") pod \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\" (UID: \"f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5\") " Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.275932 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.275953 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.275967 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.275981 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.275995 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkvdh\" (UniqueName: \"kubernetes.io/projected/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-kube-api-access-jkvdh\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.276007 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.276018 4894 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.277347 4894 scope.go:117] "RemoveContainer" containerID="09d4925167d5ad7e670ed67ebb75fd150a844f525762c0ae9cbb92bf77b00b2f" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.277748 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-kube-api-access-gl4wb" (OuterVolumeSpecName: "kube-api-access-gl4wb") pod "f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5" (UID: "f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5"). InnerVolumeSpecName "kube-api-access-gl4wb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.297864 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-config-data" (OuterVolumeSpecName: "config-data") pod "41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" (UID: "41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.298316 4894 scope.go:117] "RemoveContainer" containerID="49883e5ea70af46c31fe682fee954dba8d75f41207f6cc81ea8f694214b908c9" Dec 09 15:52:44 crc kubenswrapper[4894]: E1209 15:52:44.298920 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49883e5ea70af46c31fe682fee954dba8d75f41207f6cc81ea8f694214b908c9\": container with ID starting with 49883e5ea70af46c31fe682fee954dba8d75f41207f6cc81ea8f694214b908c9 not found: ID does not exist" containerID="49883e5ea70af46c31fe682fee954dba8d75f41207f6cc81ea8f694214b908c9" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.298957 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49883e5ea70af46c31fe682fee954dba8d75f41207f6cc81ea8f694214b908c9"} err="failed to get container status \"49883e5ea70af46c31fe682fee954dba8d75f41207f6cc81ea8f694214b908c9\": rpc error: code = NotFound desc = could not find container \"49883e5ea70af46c31fe682fee954dba8d75f41207f6cc81ea8f694214b908c9\": container with ID starting with 49883e5ea70af46c31fe682fee954dba8d75f41207f6cc81ea8f694214b908c9 not found: ID does not exist" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.298985 4894 scope.go:117] "RemoveContainer" containerID="09d4925167d5ad7e670ed67ebb75fd150a844f525762c0ae9cbb92bf77b00b2f" Dec 09 15:52:44 crc kubenswrapper[4894]: E1209 15:52:44.299283 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"09d4925167d5ad7e670ed67ebb75fd150a844f525762c0ae9cbb92bf77b00b2f\": container with ID starting with 09d4925167d5ad7e670ed67ebb75fd150a844f525762c0ae9cbb92bf77b00b2f not found: ID does not exist" containerID="09d4925167d5ad7e670ed67ebb75fd150a844f525762c0ae9cbb92bf77b00b2f" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.299316 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"09d4925167d5ad7e670ed67ebb75fd150a844f525762c0ae9cbb92bf77b00b2f"} err="failed to get container status \"09d4925167d5ad7e670ed67ebb75fd150a844f525762c0ae9cbb92bf77b00b2f\": rpc error: code = NotFound desc = could not find container \"09d4925167d5ad7e670ed67ebb75fd150a844f525762c0ae9cbb92bf77b00b2f\": container with ID starting with 09d4925167d5ad7e670ed67ebb75fd150a844f525762c0ae9cbb92bf77b00b2f not found: ID does not exist" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.299336 4894 scope.go:117] "RemoveContainer" containerID="7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.318386 4894 scope.go:117] "RemoveContainer" containerID="a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.318803 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5" (UID: "f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.318900 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5" (UID: "f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.325039 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-config" (OuterVolumeSpecName: "config") pod "f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5" (UID: "f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.326536 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5" (UID: "f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.345399 4894 scope.go:117] "RemoveContainer" containerID="b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.367741 4894 scope.go:117] "RemoveContainer" containerID="ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.379306 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.381078 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.381214 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.381276 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gl4wb\" (UniqueName: \"kubernetes.io/projected/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-kube-api-access-gl4wb\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.381343 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.381404 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.405942 4894 scope.go:117] "RemoveContainer" containerID="7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.406111 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.406309 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="6a185555-06a5-4630-8d9c-700c44af003a" containerName="nova-scheduler-scheduler" containerID="cri-o://d50a6675c092a631fc70b8d36a691aea8241c786f666ebeacf9ea9b4d56c78db" gracePeriod=30 Dec 09 15:52:44 crc kubenswrapper[4894]: E1209 15:52:44.406341 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810\": container with ID starting with 7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810 not found: ID does not exist" containerID="7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.406377 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810"} err="failed to get container status \"7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810\": rpc error: code = NotFound desc = could not find container \"7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810\": container with ID starting with 7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810 not found: ID does not exist" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.406404 4894 scope.go:117] "RemoveContainer" containerID="a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f" Dec 09 15:52:44 crc kubenswrapper[4894]: E1209 15:52:44.406961 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f\": container with ID starting with a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f not found: ID does not exist" containerID="a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.407003 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f"} err="failed to get container status \"a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f\": rpc error: code = NotFound desc = could not find container \"a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f\": container with ID starting with a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f not found: ID does not exist" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.407032 4894 scope.go:117] "RemoveContainer" containerID="b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838" Dec 09 15:52:44 crc kubenswrapper[4894]: E1209 15:52:44.407344 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838\": container with ID starting with b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838 not found: ID does not exist" containerID="b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.407379 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838"} err="failed to get container status \"b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838\": rpc error: code = NotFound desc = could not find container \"b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838\": container with ID starting with b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838 not found: ID does not exist" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.407404 4894 scope.go:117] "RemoveContainer" containerID="ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194" Dec 09 15:52:44 crc kubenswrapper[4894]: E1209 15:52:44.409147 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194\": container with ID starting with ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194 not found: ID does not exist" containerID="ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.409176 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194"} err="failed to get container status \"ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194\": rpc error: code = NotFound desc = could not find container \"ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194\": container with ID starting with ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194 not found: ID does not exist" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.409197 4894 scope.go:117] "RemoveContainer" containerID="7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.410965 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810"} err="failed to get container status \"7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810\": rpc error: code = NotFound desc = could not find container \"7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810\": container with ID starting with 7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810 not found: ID does not exist" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.411086 4894 scope.go:117] "RemoveContainer" containerID="a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.412774 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f"} err="failed to get container status \"a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f\": rpc error: code = NotFound desc = could not find container \"a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f\": container with ID starting with a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f not found: ID does not exist" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.412822 4894 scope.go:117] "RemoveContainer" containerID="b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.413439 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838"} err="failed to get container status \"b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838\": rpc error: code = NotFound desc = could not find container \"b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838\": container with ID starting with b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838 not found: ID does not exist" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.413484 4894 scope.go:117] "RemoveContainer" containerID="ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.414297 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194"} err="failed to get container status \"ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194\": rpc error: code = NotFound desc = could not find container \"ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194\": container with ID starting with ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194 not found: ID does not exist" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.414325 4894 scope.go:117] "RemoveContainer" containerID="7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.414527 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810"} err="failed to get container status \"7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810\": rpc error: code = NotFound desc = could not find container \"7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810\": container with ID starting with 7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810 not found: ID does not exist" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.414546 4894 scope.go:117] "RemoveContainer" containerID="a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.414734 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f"} err="failed to get container status \"a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f\": rpc error: code = NotFound desc = could not find container \"a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f\": container with ID starting with a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f not found: ID does not exist" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.414754 4894 scope.go:117] "RemoveContainer" containerID="b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.416374 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838"} err="failed to get container status \"b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838\": rpc error: code = NotFound desc = could not find container \"b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838\": container with ID starting with b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838 not found: ID does not exist" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.416419 4894 scope.go:117] "RemoveContainer" containerID="ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.416784 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194"} err="failed to get container status \"ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194\": rpc error: code = NotFound desc = could not find container \"ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194\": container with ID starting with ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194 not found: ID does not exist" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.416822 4894 scope.go:117] "RemoveContainer" containerID="7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.417229 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810"} err="failed to get container status \"7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810\": rpc error: code = NotFound desc = could not find container \"7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810\": container with ID starting with 7a3452686d1e9ddf4abe192d0af79b803e411a34c8d312ef2486df57fcab6810 not found: ID does not exist" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.417249 4894 scope.go:117] "RemoveContainer" containerID="a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.417459 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f"} err="failed to get container status \"a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f\": rpc error: code = NotFound desc = could not find container \"a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f\": container with ID starting with a0d373693491b9cdbcddb6f5965dde4b41841d770a6ef6e965318ecf747feb7f not found: ID does not exist" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.417477 4894 scope.go:117] "RemoveContainer" containerID="b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.417916 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838"} err="failed to get container status \"b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838\": rpc error: code = NotFound desc = could not find container \"b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838\": container with ID starting with b33da4705995743881d04aa11af0af32b7532fcbe1ff0aefe21a8ee77ed1d838 not found: ID does not exist" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.418009 4894 scope.go:117] "RemoveContainer" containerID="ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.418335 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194"} err="failed to get container status \"ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194\": rpc error: code = NotFound desc = could not find container \"ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194\": container with ID starting with ffa63abc9e86b542b05861cad0d79c35e6ea37cf311fe84b0b6a614e846fc194 not found: ID does not exist" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.425788 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.426061 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9e6f0669-9599-480e-aa0a-9b36bfd851ff" containerName="nova-api-log" containerID="cri-o://4fe547028bba0b3a911c6a4a336af0666449ceed03554255958fc031849e6587" gracePeriod=30 Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.426164 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="9e6f0669-9599-480e-aa0a-9b36bfd851ff" containerName="nova-api-api" containerID="cri-o://e6cd81a4a234df461b1ea3e319fdfcf6d16fa0a898e255df8d62a7ba0847b35a" gracePeriod=30 Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.433440 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.433752 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e97b390d-e40e-45d2-8eb7-c4ff7205e5fb" containerName="nova-metadata-log" containerID="cri-o://ef67d9cb479d3583d45a1d6dad7d1d87ad5f2a87d3cc7232eb7e7d8a438d9dbf" gracePeriod=30 Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.434654 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="e97b390d-e40e-45d2-8eb7-c4ff7205e5fb" containerName="nova-metadata-metadata" containerID="cri-o://8f81a06f5c2089ac5ad882ae5ee1988d11b7bd4006892b97fb4e4d6aaf466596" gracePeriod=30 Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.592248 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8d97cbc7-jdzrx"] Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.600674 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8d97cbc7-jdzrx"] Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.614128 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.623914 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.645023 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:52:44 crc kubenswrapper[4894]: E1209 15:52:44.645463 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5" containerName="init" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.649369 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5" containerName="init" Dec 09 15:52:44 crc kubenswrapper[4894]: E1209 15:52:44.649427 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" containerName="ceilometer-notification-agent" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.649437 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" containerName="ceilometer-notification-agent" Dec 09 15:52:44 crc kubenswrapper[4894]: E1209 15:52:44.649523 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" containerName="proxy-httpd" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.649531 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" containerName="proxy-httpd" Dec 09 15:52:44 crc kubenswrapper[4894]: E1209 15:52:44.649539 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5" containerName="dnsmasq-dns" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.649546 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5" containerName="dnsmasq-dns" Dec 09 15:52:44 crc kubenswrapper[4894]: E1209 15:52:44.649561 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95bff0ec-6b68-439a-b478-de6549452662" containerName="nova-manage" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.649570 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="95bff0ec-6b68-439a-b478-de6549452662" containerName="nova-manage" Dec 09 15:52:44 crc kubenswrapper[4894]: E1209 15:52:44.649577 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" containerName="sg-core" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.649584 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" containerName="sg-core" Dec 09 15:52:44 crc kubenswrapper[4894]: E1209 15:52:44.649596 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" containerName="ceilometer-central-agent" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.649604 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" containerName="ceilometer-central-agent" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.649968 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" containerName="ceilometer-central-agent" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.649991 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" containerName="sg-core" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.650020 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5" containerName="dnsmasq-dns" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.650030 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="95bff0ec-6b68-439a-b478-de6549452662" containerName="nova-manage" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.650039 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" containerName="proxy-httpd" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.650052 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" containerName="ceilometer-notification-agent" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.652179 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.663027 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.666603 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.667662 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.673628 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.788102 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.788164 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-run-httpd\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.788261 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-log-httpd\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.788309 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.788341 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-config-data\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.788371 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.788390 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx8l6\" (UniqueName: \"kubernetes.io/projected/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-kube-api-access-zx8l6\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.788447 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-scripts\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.890422 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.890470 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-run-httpd\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.890541 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-log-httpd\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.890588 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.890619 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-config-data\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.890656 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.890673 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx8l6\" (UniqueName: \"kubernetes.io/projected/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-kube-api-access-zx8l6\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.890712 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-scripts\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.892730 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-log-httpd\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.893041 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-run-httpd\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.894794 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-scripts\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.894811 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.894853 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.895885 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-config-data\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.897461 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.908493 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx8l6\" (UniqueName: \"kubernetes.io/projected/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-kube-api-access-zx8l6\") pod \"ceilometer-0\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " pod="openstack/ceilometer-0" Dec 09 15:52:44 crc kubenswrapper[4894]: I1209 15:52:44.998427 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 15:52:45 crc kubenswrapper[4894]: I1209 15:52:45.471201 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.125224 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53" path="/var/lib/kubelet/pods/41dcd3f4-7cc3-4ecb-bfd1-844002cb5a53/volumes" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.132669 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5" path="/var/lib/kubelet/pods/f836ddeb-882a-49e3-ab6f-4b7aa3e6b5e5/volumes" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.259390 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.283030 4894 generic.go:334] "Generic (PLEG): container finished" podID="9e6f0669-9599-480e-aa0a-9b36bfd851ff" containerID="e6cd81a4a234df461b1ea3e319fdfcf6d16fa0a898e255df8d62a7ba0847b35a" exitCode=0 Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.283061 4894 generic.go:334] "Generic (PLEG): container finished" podID="9e6f0669-9599-480e-aa0a-9b36bfd851ff" containerID="4fe547028bba0b3a911c6a4a336af0666449ceed03554255958fc031849e6587" exitCode=143 Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.283101 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9e6f0669-9599-480e-aa0a-9b36bfd851ff","Type":"ContainerDied","Data":"e6cd81a4a234df461b1ea3e319fdfcf6d16fa0a898e255df8d62a7ba0847b35a"} Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.283127 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9e6f0669-9599-480e-aa0a-9b36bfd851ff","Type":"ContainerDied","Data":"4fe547028bba0b3a911c6a4a336af0666449ceed03554255958fc031849e6587"} Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.283136 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"9e6f0669-9599-480e-aa0a-9b36bfd851ff","Type":"ContainerDied","Data":"37e638f3418ec964ac15866c8143137f24cc30c47f41979ac5530f0e43642bf5"} Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.283149 4894 scope.go:117] "RemoveContainer" containerID="e6cd81a4a234df461b1ea3e319fdfcf6d16fa0a898e255df8d62a7ba0847b35a" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.283266 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.292131 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10","Type":"ContainerStarted","Data":"92663ce5b12d2f73b12b4b15ff83d6702f0905e84c2901180df0ea041a6cb589"} Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.294613 4894 generic.go:334] "Generic (PLEG): container finished" podID="e97b390d-e40e-45d2-8eb7-c4ff7205e5fb" containerID="ef67d9cb479d3583d45a1d6dad7d1d87ad5f2a87d3cc7232eb7e7d8a438d9dbf" exitCode=143 Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.294664 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb","Type":"ContainerDied","Data":"ef67d9cb479d3583d45a1d6dad7d1d87ad5f2a87d3cc7232eb7e7d8a438d9dbf"} Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.325692 4894 scope.go:117] "RemoveContainer" containerID="4fe547028bba0b3a911c6a4a336af0666449ceed03554255958fc031849e6587" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.341954 4894 scope.go:117] "RemoveContainer" containerID="e6cd81a4a234df461b1ea3e319fdfcf6d16fa0a898e255df8d62a7ba0847b35a" Dec 09 15:52:46 crc kubenswrapper[4894]: E1209 15:52:46.342442 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e6cd81a4a234df461b1ea3e319fdfcf6d16fa0a898e255df8d62a7ba0847b35a\": container with ID starting with e6cd81a4a234df461b1ea3e319fdfcf6d16fa0a898e255df8d62a7ba0847b35a not found: ID does not exist" containerID="e6cd81a4a234df461b1ea3e319fdfcf6d16fa0a898e255df8d62a7ba0847b35a" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.342484 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6cd81a4a234df461b1ea3e319fdfcf6d16fa0a898e255df8d62a7ba0847b35a"} err="failed to get container status \"e6cd81a4a234df461b1ea3e319fdfcf6d16fa0a898e255df8d62a7ba0847b35a\": rpc error: code = NotFound desc = could not find container \"e6cd81a4a234df461b1ea3e319fdfcf6d16fa0a898e255df8d62a7ba0847b35a\": container with ID starting with e6cd81a4a234df461b1ea3e319fdfcf6d16fa0a898e255df8d62a7ba0847b35a not found: ID does not exist" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.342509 4894 scope.go:117] "RemoveContainer" containerID="4fe547028bba0b3a911c6a4a336af0666449ceed03554255958fc031849e6587" Dec 09 15:52:46 crc kubenswrapper[4894]: E1209 15:52:46.342977 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4fe547028bba0b3a911c6a4a336af0666449ceed03554255958fc031849e6587\": container with ID starting with 4fe547028bba0b3a911c6a4a336af0666449ceed03554255958fc031849e6587 not found: ID does not exist" containerID="4fe547028bba0b3a911c6a4a336af0666449ceed03554255958fc031849e6587" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.343008 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fe547028bba0b3a911c6a4a336af0666449ceed03554255958fc031849e6587"} err="failed to get container status \"4fe547028bba0b3a911c6a4a336af0666449ceed03554255958fc031849e6587\": rpc error: code = NotFound desc = could not find container \"4fe547028bba0b3a911c6a4a336af0666449ceed03554255958fc031849e6587\": container with ID starting with 4fe547028bba0b3a911c6a4a336af0666449ceed03554255958fc031849e6587 not found: ID does not exist" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.343050 4894 scope.go:117] "RemoveContainer" containerID="e6cd81a4a234df461b1ea3e319fdfcf6d16fa0a898e255df8d62a7ba0847b35a" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.343434 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e6cd81a4a234df461b1ea3e319fdfcf6d16fa0a898e255df8d62a7ba0847b35a"} err="failed to get container status \"e6cd81a4a234df461b1ea3e319fdfcf6d16fa0a898e255df8d62a7ba0847b35a\": rpc error: code = NotFound desc = could not find container \"e6cd81a4a234df461b1ea3e319fdfcf6d16fa0a898e255df8d62a7ba0847b35a\": container with ID starting with e6cd81a4a234df461b1ea3e319fdfcf6d16fa0a898e255df8d62a7ba0847b35a not found: ID does not exist" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.343453 4894 scope.go:117] "RemoveContainer" containerID="4fe547028bba0b3a911c6a4a336af0666449ceed03554255958fc031849e6587" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.343760 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4fe547028bba0b3a911c6a4a336af0666449ceed03554255958fc031849e6587"} err="failed to get container status \"4fe547028bba0b3a911c6a4a336af0666449ceed03554255958fc031849e6587\": rpc error: code = NotFound desc = could not find container \"4fe547028bba0b3a911c6a4a336af0666449ceed03554255958fc031849e6587\": container with ID starting with 4fe547028bba0b3a911c6a4a336af0666449ceed03554255958fc031849e6587 not found: ID does not exist" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.428583 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-internal-tls-certs\") pod \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.428862 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xxlzn\" (UniqueName: \"kubernetes.io/projected/9e6f0669-9599-480e-aa0a-9b36bfd851ff-kube-api-access-xxlzn\") pod \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.428987 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-config-data\") pod \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.429122 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e6f0669-9599-480e-aa0a-9b36bfd851ff-logs\") pod \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.429306 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-public-tls-certs\") pod \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.429522 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9e6f0669-9599-480e-aa0a-9b36bfd851ff-logs" (OuterVolumeSpecName: "logs") pod "9e6f0669-9599-480e-aa0a-9b36bfd851ff" (UID: "9e6f0669-9599-480e-aa0a-9b36bfd851ff"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.429654 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-combined-ca-bundle\") pod \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\" (UID: \"9e6f0669-9599-480e-aa0a-9b36bfd851ff\") " Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.430461 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9e6f0669-9599-480e-aa0a-9b36bfd851ff-logs\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.432911 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e6f0669-9599-480e-aa0a-9b36bfd851ff-kube-api-access-xxlzn" (OuterVolumeSpecName: "kube-api-access-xxlzn") pod "9e6f0669-9599-480e-aa0a-9b36bfd851ff" (UID: "9e6f0669-9599-480e-aa0a-9b36bfd851ff"). InnerVolumeSpecName "kube-api-access-xxlzn". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.455571 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9e6f0669-9599-480e-aa0a-9b36bfd851ff" (UID: "9e6f0669-9599-480e-aa0a-9b36bfd851ff"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.458217 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-config-data" (OuterVolumeSpecName: "config-data") pod "9e6f0669-9599-480e-aa0a-9b36bfd851ff" (UID: "9e6f0669-9599-480e-aa0a-9b36bfd851ff"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.482799 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "9e6f0669-9599-480e-aa0a-9b36bfd851ff" (UID: "9e6f0669-9599-480e-aa0a-9b36bfd851ff"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.490083 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "9e6f0669-9599-480e-aa0a-9b36bfd851ff" (UID: "9e6f0669-9599-480e-aa0a-9b36bfd851ff"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.532290 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.532335 4894 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.532348 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xxlzn\" (UniqueName: \"kubernetes.io/projected/9e6f0669-9599-480e-aa0a-9b36bfd851ff-kube-api-access-xxlzn\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.532363 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.532374 4894 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/9e6f0669-9599-480e-aa0a-9b36bfd851ff-public-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.614897 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.623214 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.675007 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Dec 09 15:52:46 crc kubenswrapper[4894]: E1209 15:52:46.675420 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6f0669-9599-480e-aa0a-9b36bfd851ff" containerName="nova-api-log" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.675437 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6f0669-9599-480e-aa0a-9b36bfd851ff" containerName="nova-api-log" Dec 09 15:52:46 crc kubenswrapper[4894]: E1209 15:52:46.675477 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6f0669-9599-480e-aa0a-9b36bfd851ff" containerName="nova-api-api" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.675485 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6f0669-9599-480e-aa0a-9b36bfd851ff" containerName="nova-api-api" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.675695 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e6f0669-9599-480e-aa0a-9b36bfd851ff" containerName="nova-api-log" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.675723 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e6f0669-9599-480e-aa0a-9b36bfd851ff" containerName="nova-api-api" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.676716 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.681060 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.682099 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.682225 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.687148 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.759773 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cb294ea-866f-4569-8f83-c9567f81dfd6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2cb294ea-866f-4569-8f83-c9567f81dfd6\") " pod="openstack/nova-api-0" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.759835 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2cb294ea-866f-4569-8f83-c9567f81dfd6-logs\") pod \"nova-api-0\" (UID: \"2cb294ea-866f-4569-8f83-c9567f81dfd6\") " pod="openstack/nova-api-0" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.759875 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cb294ea-866f-4569-8f83-c9567f81dfd6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2cb294ea-866f-4569-8f83-c9567f81dfd6\") " pod="openstack/nova-api-0" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.759949 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cb294ea-866f-4569-8f83-c9567f81dfd6-public-tls-certs\") pod \"nova-api-0\" (UID: \"2cb294ea-866f-4569-8f83-c9567f81dfd6\") " pod="openstack/nova-api-0" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.759966 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29q26\" (UniqueName: \"kubernetes.io/projected/2cb294ea-866f-4569-8f83-c9567f81dfd6-kube-api-access-29q26\") pod \"nova-api-0\" (UID: \"2cb294ea-866f-4569-8f83-c9567f81dfd6\") " pod="openstack/nova-api-0" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.759986 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cb294ea-866f-4569-8f83-c9567f81dfd6-config-data\") pod \"nova-api-0\" (UID: \"2cb294ea-866f-4569-8f83-c9567f81dfd6\") " pod="openstack/nova-api-0" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.860959 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cb294ea-866f-4569-8f83-c9567f81dfd6-public-tls-certs\") pod \"nova-api-0\" (UID: \"2cb294ea-866f-4569-8f83-c9567f81dfd6\") " pod="openstack/nova-api-0" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.860999 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29q26\" (UniqueName: \"kubernetes.io/projected/2cb294ea-866f-4569-8f83-c9567f81dfd6-kube-api-access-29q26\") pod \"nova-api-0\" (UID: \"2cb294ea-866f-4569-8f83-c9567f81dfd6\") " pod="openstack/nova-api-0" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.861028 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cb294ea-866f-4569-8f83-c9567f81dfd6-config-data\") pod \"nova-api-0\" (UID: \"2cb294ea-866f-4569-8f83-c9567f81dfd6\") " pod="openstack/nova-api-0" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.861114 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cb294ea-866f-4569-8f83-c9567f81dfd6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2cb294ea-866f-4569-8f83-c9567f81dfd6\") " pod="openstack/nova-api-0" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.861150 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2cb294ea-866f-4569-8f83-c9567f81dfd6-logs\") pod \"nova-api-0\" (UID: \"2cb294ea-866f-4569-8f83-c9567f81dfd6\") " pod="openstack/nova-api-0" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.861198 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cb294ea-866f-4569-8f83-c9567f81dfd6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2cb294ea-866f-4569-8f83-c9567f81dfd6\") " pod="openstack/nova-api-0" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.861969 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/2cb294ea-866f-4569-8f83-c9567f81dfd6-logs\") pod \"nova-api-0\" (UID: \"2cb294ea-866f-4569-8f83-c9567f81dfd6\") " pod="openstack/nova-api-0" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.865479 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2cb294ea-866f-4569-8f83-c9567f81dfd6-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"2cb294ea-866f-4569-8f83-c9567f81dfd6\") " pod="openstack/nova-api-0" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.866004 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2cb294ea-866f-4569-8f83-c9567f81dfd6-config-data\") pod \"nova-api-0\" (UID: \"2cb294ea-866f-4569-8f83-c9567f81dfd6\") " pod="openstack/nova-api-0" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.866809 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cb294ea-866f-4569-8f83-c9567f81dfd6-internal-tls-certs\") pod \"nova-api-0\" (UID: \"2cb294ea-866f-4569-8f83-c9567f81dfd6\") " pod="openstack/nova-api-0" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.876303 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/2cb294ea-866f-4569-8f83-c9567f81dfd6-public-tls-certs\") pod \"nova-api-0\" (UID: \"2cb294ea-866f-4569-8f83-c9567f81dfd6\") " pod="openstack/nova-api-0" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.877324 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29q26\" (UniqueName: \"kubernetes.io/projected/2cb294ea-866f-4569-8f83-c9567f81dfd6-kube-api-access-29q26\") pod \"nova-api-0\" (UID: \"2cb294ea-866f-4569-8f83-c9567f81dfd6\") " pod="openstack/nova-api-0" Dec 09 15:52:46 crc kubenswrapper[4894]: I1209 15:52:46.999116 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Dec 09 15:52:47 crc kubenswrapper[4894]: I1209 15:52:47.306298 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10","Type":"ContainerStarted","Data":"c72272b99978f0df125d44820f44392dfc1dc62bb9afde6b8726332b756ca55e"} Dec 09 15:52:47 crc kubenswrapper[4894]: I1209 15:52:47.424172 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Dec 09 15:52:47 crc kubenswrapper[4894]: I1209 15:52:47.574159 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="e97b390d-e40e-45d2-8eb7-c4ff7205e5fb" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.184:8775/\": read tcp 10.217.0.2:39534->10.217.0.184:8775: read: connection reset by peer" Dec 09 15:52:47 crc kubenswrapper[4894]: I1209 15:52:47.574162 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="e97b390d-e40e-45d2-8eb7-c4ff7205e5fb" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.184:8775/\": read tcp 10.217.0.2:39528->10.217.0.184:8775: read: connection reset by peer" Dec 09 15:52:48 crc kubenswrapper[4894]: I1209 15:52:48.116809 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e6f0669-9599-480e-aa0a-9b36bfd851ff" path="/var/lib/kubelet/pods/9e6f0669-9599-480e-aa0a-9b36bfd851ff/volumes" Dec 09 15:52:48 crc kubenswrapper[4894]: I1209 15:52:48.318912 4894 generic.go:334] "Generic (PLEG): container finished" podID="e97b390d-e40e-45d2-8eb7-c4ff7205e5fb" containerID="8f81a06f5c2089ac5ad882ae5ee1988d11b7bd4006892b97fb4e4d6aaf466596" exitCode=0 Dec 09 15:52:48 crc kubenswrapper[4894]: I1209 15:52:48.318992 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb","Type":"ContainerDied","Data":"8f81a06f5c2089ac5ad882ae5ee1988d11b7bd4006892b97fb4e4d6aaf466596"} Dec 09 15:52:48 crc kubenswrapper[4894]: I1209 15:52:48.320484 4894 generic.go:334] "Generic (PLEG): container finished" podID="6a185555-06a5-4630-8d9c-700c44af003a" containerID="d50a6675c092a631fc70b8d36a691aea8241c786f666ebeacf9ea9b4d56c78db" exitCode=0 Dec 09 15:52:48 crc kubenswrapper[4894]: I1209 15:52:48.320543 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6a185555-06a5-4630-8d9c-700c44af003a","Type":"ContainerDied","Data":"d50a6675c092a631fc70b8d36a691aea8241c786f666ebeacf9ea9b4d56c78db"} Dec 09 15:52:48 crc kubenswrapper[4894]: I1209 15:52:48.322026 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2cb294ea-866f-4569-8f83-c9567f81dfd6","Type":"ContainerStarted","Data":"1bcc5f020ca007ad9af0f2f864a0eef011d1346fda30d2d88003130560057028"} Dec 09 15:52:48 crc kubenswrapper[4894]: I1209 15:52:48.322052 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2cb294ea-866f-4569-8f83-c9567f81dfd6","Type":"ContainerStarted","Data":"8055ffa00b8feda6ee24e3c1e8eeb18c500779a846b122dffd1a9e673010bad8"} Dec 09 15:52:48 crc kubenswrapper[4894]: E1209 15:52:48.335820 4894 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d50a6675c092a631fc70b8d36a691aea8241c786f666ebeacf9ea9b4d56c78db is running failed: container process not found" containerID="d50a6675c092a631fc70b8d36a691aea8241c786f666ebeacf9ea9b4d56c78db" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 15:52:48 crc kubenswrapper[4894]: E1209 15:52:48.336183 4894 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d50a6675c092a631fc70b8d36a691aea8241c786f666ebeacf9ea9b4d56c78db is running failed: container process not found" containerID="d50a6675c092a631fc70b8d36a691aea8241c786f666ebeacf9ea9b4d56c78db" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 15:52:48 crc kubenswrapper[4894]: E1209 15:52:48.336480 4894 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d50a6675c092a631fc70b8d36a691aea8241c786f666ebeacf9ea9b4d56c78db is running failed: container process not found" containerID="d50a6675c092a631fc70b8d36a691aea8241c786f666ebeacf9ea9b4d56c78db" cmd=["/usr/bin/pgrep","-r","DRST","nova-scheduler"] Dec 09 15:52:48 crc kubenswrapper[4894]: E1209 15:52:48.336508 4894 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d50a6675c092a631fc70b8d36a691aea8241c786f666ebeacf9ea9b4d56c78db is running failed: container process not found" probeType="Readiness" pod="openstack/nova-scheduler-0" podUID="6a185555-06a5-4630-8d9c-700c44af003a" containerName="nova-scheduler-scheduler" Dec 09 15:52:48 crc kubenswrapper[4894]: I1209 15:52:48.886519 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.003769 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a185555-06a5-4630-8d9c-700c44af003a-combined-ca-bundle\") pod \"6a185555-06a5-4630-8d9c-700c44af003a\" (UID: \"6a185555-06a5-4630-8d9c-700c44af003a\") " Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.004227 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a185555-06a5-4630-8d9c-700c44af003a-config-data\") pod \"6a185555-06a5-4630-8d9c-700c44af003a\" (UID: \"6a185555-06a5-4630-8d9c-700c44af003a\") " Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.004438 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-74p29\" (UniqueName: \"kubernetes.io/projected/6a185555-06a5-4630-8d9c-700c44af003a-kube-api-access-74p29\") pod \"6a185555-06a5-4630-8d9c-700c44af003a\" (UID: \"6a185555-06a5-4630-8d9c-700c44af003a\") " Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.013180 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6a185555-06a5-4630-8d9c-700c44af003a-kube-api-access-74p29" (OuterVolumeSpecName: "kube-api-access-74p29") pod "6a185555-06a5-4630-8d9c-700c44af003a" (UID: "6a185555-06a5-4630-8d9c-700c44af003a"). InnerVolumeSpecName "kube-api-access-74p29". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.038378 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a185555-06a5-4630-8d9c-700c44af003a-config-data" (OuterVolumeSpecName: "config-data") pod "6a185555-06a5-4630-8d9c-700c44af003a" (UID: "6a185555-06a5-4630-8d9c-700c44af003a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.064002 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6a185555-06a5-4630-8d9c-700c44af003a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6a185555-06a5-4630-8d9c-700c44af003a" (UID: "6a185555-06a5-4630-8d9c-700c44af003a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.122518 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-74p29\" (UniqueName: \"kubernetes.io/projected/6a185555-06a5-4630-8d9c-700c44af003a-kube-api-access-74p29\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.122542 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6a185555-06a5-4630-8d9c-700c44af003a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.122551 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6a185555-06a5-4630-8d9c-700c44af003a-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.209209 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.334423 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-logs\") pod \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\" (UID: \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\") " Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.334587 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-nova-metadata-tls-certs\") pod \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\" (UID: \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\") " Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.334622 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-config-data\") pod \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\" (UID: \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\") " Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.334695 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-combined-ca-bundle\") pod \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\" (UID: \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\") " Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.334734 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9dnw2\" (UniqueName: \"kubernetes.io/projected/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-kube-api-access-9dnw2\") pod \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\" (UID: \"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb\") " Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.338410 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-logs" (OuterVolumeSpecName: "logs") pod "e97b390d-e40e-45d2-8eb7-c4ff7205e5fb" (UID: "e97b390d-e40e-45d2-8eb7-c4ff7205e5fb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.345250 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-kube-api-access-9dnw2" (OuterVolumeSpecName: "kube-api-access-9dnw2") pod "e97b390d-e40e-45d2-8eb7-c4ff7205e5fb" (UID: "e97b390d-e40e-45d2-8eb7-c4ff7205e5fb"). InnerVolumeSpecName "kube-api-access-9dnw2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.368477 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"e97b390d-e40e-45d2-8eb7-c4ff7205e5fb","Type":"ContainerDied","Data":"62493d8e3b705c3e5b04425414000d458b9741f791db79fa72c7af19236476e3"} Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.368526 4894 scope.go:117] "RemoveContainer" containerID="8f81a06f5c2089ac5ad882ae5ee1988d11b7bd4006892b97fb4e4d6aaf466596" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.368674 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.377101 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.377038 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"6a185555-06a5-4630-8d9c-700c44af003a","Type":"ContainerDied","Data":"767349dea3cea5a4aef57e7cfcc98adcfb0a2efb9c62e85af517260abde611cb"} Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.377908 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-config-data" (OuterVolumeSpecName: "config-data") pod "e97b390d-e40e-45d2-8eb7-c4ff7205e5fb" (UID: "e97b390d-e40e-45d2-8eb7-c4ff7205e5fb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.418579 4894 scope.go:117] "RemoveContainer" containerID="ef67d9cb479d3583d45a1d6dad7d1d87ad5f2a87d3cc7232eb7e7d8a438d9dbf" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.420940 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "e97b390d-e40e-45d2-8eb7-c4ff7205e5fb" (UID: "e97b390d-e40e-45d2-8eb7-c4ff7205e5fb"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.421119 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"2cb294ea-866f-4569-8f83-c9567f81dfd6","Type":"ContainerStarted","Data":"0a1220a9665e916d05b3507a4fd801ba4b48bfdbbeb562b29e02e3396b913699"} Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.438591 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-logs\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.438662 4894 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.438676 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.438712 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9dnw2\" (UniqueName: \"kubernetes.io/projected/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-kube-api-access-9dnw2\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.457518 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e97b390d-e40e-45d2-8eb7-c4ff7205e5fb" (UID: "e97b390d-e40e-45d2-8eb7-c4ff7205e5fb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.464913 4894 scope.go:117] "RemoveContainer" containerID="d50a6675c092a631fc70b8d36a691aea8241c786f666ebeacf9ea9b4d56c78db" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.486045 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.509270 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.529142 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 15:52:49 crc kubenswrapper[4894]: E1209 15:52:49.529784 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6a185555-06a5-4630-8d9c-700c44af003a" containerName="nova-scheduler-scheduler" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.529817 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="6a185555-06a5-4630-8d9c-700c44af003a" containerName="nova-scheduler-scheduler" Dec 09 15:52:49 crc kubenswrapper[4894]: E1209 15:52:49.529847 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e97b390d-e40e-45d2-8eb7-c4ff7205e5fb" containerName="nova-metadata-log" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.529853 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e97b390d-e40e-45d2-8eb7-c4ff7205e5fb" containerName="nova-metadata-log" Dec 09 15:52:49 crc kubenswrapper[4894]: E1209 15:52:49.529864 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e97b390d-e40e-45d2-8eb7-c4ff7205e5fb" containerName="nova-metadata-metadata" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.529870 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e97b390d-e40e-45d2-8eb7-c4ff7205e5fb" containerName="nova-metadata-metadata" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.530061 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e97b390d-e40e-45d2-8eb7-c4ff7205e5fb" containerName="nova-metadata-metadata" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.530080 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="6a185555-06a5-4630-8d9c-700c44af003a" containerName="nova-scheduler-scheduler" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.530089 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e97b390d-e40e-45d2-8eb7-c4ff7205e5fb" containerName="nova-metadata-log" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.530913 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.534448 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.540886 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.544973 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.555069 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.555045917 podStartE2EDuration="3.555045917s" podCreationTimestamp="2025-12-09 15:52:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:52:49.499432566 +0000 UTC m=+1263.818643235" watchObservedRunningTime="2025-12-09 15:52:49.555045917 +0000 UTC m=+1263.874256586" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.644060 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a8b0e9c-1018-402d-9291-94588611638e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2a8b0e9c-1018-402d-9291-94588611638e\") " pod="openstack/nova-scheduler-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.644208 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a8b0e9c-1018-402d-9291-94588611638e-config-data\") pod \"nova-scheduler-0\" (UID: \"2a8b0e9c-1018-402d-9291-94588611638e\") " pod="openstack/nova-scheduler-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.644258 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbdzw\" (UniqueName: \"kubernetes.io/projected/2a8b0e9c-1018-402d-9291-94588611638e-kube-api-access-jbdzw\") pod \"nova-scheduler-0\" (UID: \"2a8b0e9c-1018-402d-9291-94588611638e\") " pod="openstack/nova-scheduler-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.707150 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.720353 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.747093 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.749016 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a8b0e9c-1018-402d-9291-94588611638e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2a8b0e9c-1018-402d-9291-94588611638e\") " pod="openstack/nova-scheduler-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.749326 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a8b0e9c-1018-402d-9291-94588611638e-config-data\") pod \"nova-scheduler-0\" (UID: \"2a8b0e9c-1018-402d-9291-94588611638e\") " pod="openstack/nova-scheduler-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.749375 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbdzw\" (UniqueName: \"kubernetes.io/projected/2a8b0e9c-1018-402d-9291-94588611638e-kube-api-access-jbdzw\") pod \"nova-scheduler-0\" (UID: \"2a8b0e9c-1018-402d-9291-94588611638e\") " pod="openstack/nova-scheduler-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.750959 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.753853 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.753915 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2a8b0e9c-1018-402d-9291-94588611638e-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"2a8b0e9c-1018-402d-9291-94588611638e\") " pod="openstack/nova-scheduler-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.755051 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2a8b0e9c-1018-402d-9291-94588611638e-config-data\") pod \"nova-scheduler-0\" (UID: \"2a8b0e9c-1018-402d-9291-94588611638e\") " pod="openstack/nova-scheduler-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.757956 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.779838 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.784905 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbdzw\" (UniqueName: \"kubernetes.io/projected/2a8b0e9c-1018-402d-9291-94588611638e-kube-api-access-jbdzw\") pod \"nova-scheduler-0\" (UID: \"2a8b0e9c-1018-402d-9291-94588611638e\") " pod="openstack/nova-scheduler-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.851392 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c8002b-4901-4a76-8035-66620904f04b-logs\") pod \"nova-metadata-0\" (UID: \"a1c8002b-4901-4a76-8035-66620904f04b\") " pod="openstack/nova-metadata-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.851678 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1c8002b-4901-4a76-8035-66620904f04b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a1c8002b-4901-4a76-8035-66620904f04b\") " pod="openstack/nova-metadata-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.851942 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1c8002b-4901-4a76-8035-66620904f04b-config-data\") pod \"nova-metadata-0\" (UID: \"a1c8002b-4901-4a76-8035-66620904f04b\") " pod="openstack/nova-metadata-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.852026 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1c8002b-4901-4a76-8035-66620904f04b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a1c8002b-4901-4a76-8035-66620904f04b\") " pod="openstack/nova-metadata-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.852074 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m5ltg\" (UniqueName: \"kubernetes.io/projected/a1c8002b-4901-4a76-8035-66620904f04b-kube-api-access-m5ltg\") pod \"nova-metadata-0\" (UID: \"a1c8002b-4901-4a76-8035-66620904f04b\") " pod="openstack/nova-metadata-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.892826 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.953669 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1c8002b-4901-4a76-8035-66620904f04b-config-data\") pod \"nova-metadata-0\" (UID: \"a1c8002b-4901-4a76-8035-66620904f04b\") " pod="openstack/nova-metadata-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.953743 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1c8002b-4901-4a76-8035-66620904f04b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a1c8002b-4901-4a76-8035-66620904f04b\") " pod="openstack/nova-metadata-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.953785 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m5ltg\" (UniqueName: \"kubernetes.io/projected/a1c8002b-4901-4a76-8035-66620904f04b-kube-api-access-m5ltg\") pod \"nova-metadata-0\" (UID: \"a1c8002b-4901-4a76-8035-66620904f04b\") " pod="openstack/nova-metadata-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.953835 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c8002b-4901-4a76-8035-66620904f04b-logs\") pod \"nova-metadata-0\" (UID: \"a1c8002b-4901-4a76-8035-66620904f04b\") " pod="openstack/nova-metadata-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.953886 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1c8002b-4901-4a76-8035-66620904f04b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a1c8002b-4901-4a76-8035-66620904f04b\") " pod="openstack/nova-metadata-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.956998 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a1c8002b-4901-4a76-8035-66620904f04b-logs\") pod \"nova-metadata-0\" (UID: \"a1c8002b-4901-4a76-8035-66620904f04b\") " pod="openstack/nova-metadata-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.957859 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a1c8002b-4901-4a76-8035-66620904f04b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"a1c8002b-4901-4a76-8035-66620904f04b\") " pod="openstack/nova-metadata-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.958319 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/a1c8002b-4901-4a76-8035-66620904f04b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"a1c8002b-4901-4a76-8035-66620904f04b\") " pod="openstack/nova-metadata-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.960317 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a1c8002b-4901-4a76-8035-66620904f04b-config-data\") pod \"nova-metadata-0\" (UID: \"a1c8002b-4901-4a76-8035-66620904f04b\") " pod="openstack/nova-metadata-0" Dec 09 15:52:49 crc kubenswrapper[4894]: I1209 15:52:49.974245 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m5ltg\" (UniqueName: \"kubernetes.io/projected/a1c8002b-4901-4a76-8035-66620904f04b-kube-api-access-m5ltg\") pod \"nova-metadata-0\" (UID: \"a1c8002b-4901-4a76-8035-66620904f04b\") " pod="openstack/nova-metadata-0" Dec 09 15:52:50 crc kubenswrapper[4894]: I1209 15:52:50.120970 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6a185555-06a5-4630-8d9c-700c44af003a" path="/var/lib/kubelet/pods/6a185555-06a5-4630-8d9c-700c44af003a/volumes" Dec 09 15:52:50 crc kubenswrapper[4894]: I1209 15:52:50.122337 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e97b390d-e40e-45d2-8eb7-c4ff7205e5fb" path="/var/lib/kubelet/pods/e97b390d-e40e-45d2-8eb7-c4ff7205e5fb/volumes" Dec 09 15:52:50 crc kubenswrapper[4894]: I1209 15:52:50.146331 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Dec 09 15:52:50 crc kubenswrapper[4894]: I1209 15:52:50.334931 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Dec 09 15:52:50 crc kubenswrapper[4894]: W1209 15:52:50.338036 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a8b0e9c_1018_402d_9291_94588611638e.slice/crio-401ca4a29c4872a3087e892c43b72d29c574e08814cb733ea3634fde52cf9517 WatchSource:0}: Error finding container 401ca4a29c4872a3087e892c43b72d29c574e08814cb733ea3634fde52cf9517: Status 404 returned error can't find the container with id 401ca4a29c4872a3087e892c43b72d29c574e08814cb733ea3634fde52cf9517 Dec 09 15:52:50 crc kubenswrapper[4894]: I1209 15:52:50.433889 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2a8b0e9c-1018-402d-9291-94588611638e","Type":"ContainerStarted","Data":"401ca4a29c4872a3087e892c43b72d29c574e08814cb733ea3634fde52cf9517"} Dec 09 15:52:50 crc kubenswrapper[4894]: I1209 15:52:50.439975 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10","Type":"ContainerStarted","Data":"4c5103ffb3f6284c1b3040424af759fb01f8cc7149c20c54a01d1c212d3091b8"} Dec 09 15:52:50 crc kubenswrapper[4894]: I1209 15:52:50.644063 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Dec 09 15:52:51 crc kubenswrapper[4894]: I1209 15:52:51.463116 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"2a8b0e9c-1018-402d-9291-94588611638e","Type":"ContainerStarted","Data":"5fbd605e9699b66bf1fd6a5b6b8ba3e8eae9a505afd3129e0cb5072ae9f08cf8"} Dec 09 15:52:51 crc kubenswrapper[4894]: I1209 15:52:51.466445 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10","Type":"ContainerStarted","Data":"97f3046a73e8242e7c85763c02199b24c68736ef49368531c0931f58272d480e"} Dec 09 15:52:51 crc kubenswrapper[4894]: I1209 15:52:51.468286 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1c8002b-4901-4a76-8035-66620904f04b","Type":"ContainerStarted","Data":"8df7dfc518f01087d1c12bb6765bee28570167b1611eadd57b3d2111854b8cda"} Dec 09 15:52:51 crc kubenswrapper[4894]: I1209 15:52:51.468312 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1c8002b-4901-4a76-8035-66620904f04b","Type":"ContainerStarted","Data":"e3b92a5f0ea908bad3af008191a1a473c2c2885e1859216678ca01c8460cabcf"} Dec 09 15:52:51 crc kubenswrapper[4894]: I1209 15:52:51.468321 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"a1c8002b-4901-4a76-8035-66620904f04b","Type":"ContainerStarted","Data":"69ca1e84e22b63271877ae8328b1a3a6b0b06155e940e33e7a57a6141f535d6d"} Dec 09 15:52:51 crc kubenswrapper[4894]: I1209 15:52:51.491912 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.491890389 podStartE2EDuration="2.491890389s" podCreationTimestamp="2025-12-09 15:52:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:52:51.485822795 +0000 UTC m=+1265.805033484" watchObservedRunningTime="2025-12-09 15:52:51.491890389 +0000 UTC m=+1265.811101058" Dec 09 15:52:51 crc kubenswrapper[4894]: I1209 15:52:51.502590 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.502564707 podStartE2EDuration="2.502564707s" podCreationTimestamp="2025-12-09 15:52:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:52:51.501378635 +0000 UTC m=+1265.820589314" watchObservedRunningTime="2025-12-09 15:52:51.502564707 +0000 UTC m=+1265.821775376" Dec 09 15:52:54 crc kubenswrapper[4894]: I1209 15:52:54.893741 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Dec 09 15:52:55 crc kubenswrapper[4894]: I1209 15:52:55.147133 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 09 15:52:55 crc kubenswrapper[4894]: I1209 15:52:55.147228 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Dec 09 15:52:55 crc kubenswrapper[4894]: I1209 15:52:55.507561 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10","Type":"ContainerStarted","Data":"e2832eefa685d75df52c333784a4e0e53123f9dc331b58ea1d921d9235ece700"} Dec 09 15:52:56 crc kubenswrapper[4894]: I1209 15:52:56.515466 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 15:52:56 crc kubenswrapper[4894]: I1209 15:52:56.546911 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.234858322 podStartE2EDuration="12.54689293s" podCreationTimestamp="2025-12-09 15:52:44 +0000 UTC" firstStartedPulling="2025-12-09 15:52:45.472567568 +0000 UTC m=+1259.791778237" lastFinishedPulling="2025-12-09 15:52:53.784602176 +0000 UTC m=+1268.103812845" observedRunningTime="2025-12-09 15:52:56.541938317 +0000 UTC m=+1270.861149006" watchObservedRunningTime="2025-12-09 15:52:56.54689293 +0000 UTC m=+1270.866103599" Dec 09 15:52:57 crc kubenswrapper[4894]: I1209 15:52:56.999984 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 15:52:57 crc kubenswrapper[4894]: I1209 15:52:57.000432 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Dec 09 15:52:58 crc kubenswrapper[4894]: I1209 15:52:58.013826 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2cb294ea-866f-4569-8f83-c9567f81dfd6" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.195:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 15:52:58 crc kubenswrapper[4894]: I1209 15:52:58.014017 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="2cb294ea-866f-4569-8f83-c9567f81dfd6" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.195:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 15:52:59 crc kubenswrapper[4894]: I1209 15:52:59.894194 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Dec 09 15:52:59 crc kubenswrapper[4894]: I1209 15:52:59.925457 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Dec 09 15:53:00 crc kubenswrapper[4894]: I1209 15:53:00.147127 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 09 15:53:00 crc kubenswrapper[4894]: I1209 15:53:00.147302 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Dec 09 15:53:00 crc kubenswrapper[4894]: I1209 15:53:00.579746 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Dec 09 15:53:01 crc kubenswrapper[4894]: I1209 15:53:01.161880 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a1c8002b-4901-4a76-8035-66620904f04b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 15:53:01 crc kubenswrapper[4894]: I1209 15:53:01.161875 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="a1c8002b-4901-4a76-8035-66620904f04b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.197:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Dec 09 15:53:07 crc kubenswrapper[4894]: I1209 15:53:07.007131 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 09 15:53:07 crc kubenswrapper[4894]: I1209 15:53:07.008353 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 09 15:53:07 crc kubenswrapper[4894]: I1209 15:53:07.008771 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Dec 09 15:53:07 crc kubenswrapper[4894]: I1209 15:53:07.015265 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 09 15:53:07 crc kubenswrapper[4894]: I1209 15:53:07.623961 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Dec 09 15:53:07 crc kubenswrapper[4894]: I1209 15:53:07.631275 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Dec 09 15:53:10 crc kubenswrapper[4894]: I1209 15:53:10.151545 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 09 15:53:10 crc kubenswrapper[4894]: I1209 15:53:10.153832 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Dec 09 15:53:10 crc kubenswrapper[4894]: I1209 15:53:10.157782 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 09 15:53:10 crc kubenswrapper[4894]: I1209 15:53:10.665428 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Dec 09 15:53:12 crc kubenswrapper[4894]: I1209 15:53:12.276734 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:53:12 crc kubenswrapper[4894]: I1209 15:53:12.278084 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:53:15 crc kubenswrapper[4894]: I1209 15:53:15.006952 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 09 15:53:24 crc kubenswrapper[4894]: I1209 15:53:24.249848 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 15:53:25 crc kubenswrapper[4894]: I1209 15:53:25.753156 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 15:53:28 crc kubenswrapper[4894]: I1209 15:53:28.405468 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="493425e2-b4c7-4a1b-af05-44b4f44bd2b3" containerName="rabbitmq" containerID="cri-o://521f1e5dfce4a639e4244abf43400cbc86ec270828894f24b2104650d695e1f3" gracePeriod=604796 Dec 09 15:53:29 crc kubenswrapper[4894]: I1209 15:53:29.966481 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="44fdaebd-a96c-47f9-a421-a0984f78972e" containerName="rabbitmq" containerID="cri-o://0ebd4954c8e209f4c21212c42fd48068abec230259a29738503745230bd00e15" gracePeriod=604796 Dec 09 15:53:34 crc kubenswrapper[4894]: I1209 15:53:34.875970 4894 generic.go:334] "Generic (PLEG): container finished" podID="493425e2-b4c7-4a1b-af05-44b4f44bd2b3" containerID="521f1e5dfce4a639e4244abf43400cbc86ec270828894f24b2104650d695e1f3" exitCode=0 Dec 09 15:53:34 crc kubenswrapper[4894]: I1209 15:53:34.876044 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"493425e2-b4c7-4a1b-af05-44b4f44bd2b3","Type":"ContainerDied","Data":"521f1e5dfce4a639e4244abf43400cbc86ec270828894f24b2104650d695e1f3"} Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.045297 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.202043 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b22pt\" (UniqueName: \"kubernetes.io/projected/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-kube-api-access-b22pt\") pod \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.202483 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-erlang-cookie-secret\") pod \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.202833 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-server-conf\") pod \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.202876 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-plugins-conf\") pod \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.202893 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-pod-info\") pod \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.202910 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-confd\") pod \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.202934 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-erlang-cookie\") pod \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.202970 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-plugins\") pod \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.202998 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-config-data\") pod \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.203018 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-tls\") pod \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.203082 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\" (UID: \"493425e2-b4c7-4a1b-af05-44b4f44bd2b3\") " Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.203404 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "493425e2-b4c7-4a1b-af05-44b4f44bd2b3" (UID: "493425e2-b4c7-4a1b-af05-44b4f44bd2b3"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.205434 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "493425e2-b4c7-4a1b-af05-44b4f44bd2b3" (UID: "493425e2-b4c7-4a1b-af05-44b4f44bd2b3"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.218902 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "persistence") pod "493425e2-b4c7-4a1b-af05-44b4f44bd2b3" (UID: "493425e2-b4c7-4a1b-af05-44b4f44bd2b3"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.221568 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-kube-api-access-b22pt" (OuterVolumeSpecName: "kube-api-access-b22pt") pod "493425e2-b4c7-4a1b-af05-44b4f44bd2b3" (UID: "493425e2-b4c7-4a1b-af05-44b4f44bd2b3"). InnerVolumeSpecName "kube-api-access-b22pt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.222256 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "493425e2-b4c7-4a1b-af05-44b4f44bd2b3" (UID: "493425e2-b4c7-4a1b-af05-44b4f44bd2b3"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.222290 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "493425e2-b4c7-4a1b-af05-44b4f44bd2b3" (UID: "493425e2-b4c7-4a1b-af05-44b4f44bd2b3"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.222506 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "493425e2-b4c7-4a1b-af05-44b4f44bd2b3" (UID: "493425e2-b4c7-4a1b-af05-44b4f44bd2b3"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.224983 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-pod-info" (OuterVolumeSpecName: "pod-info") pod "493425e2-b4c7-4a1b-af05-44b4f44bd2b3" (UID: "493425e2-b4c7-4a1b-af05-44b4f44bd2b3"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.255354 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-config-data" (OuterVolumeSpecName: "config-data") pod "493425e2-b4c7-4a1b-af05-44b4f44bd2b3" (UID: "493425e2-b4c7-4a1b-af05-44b4f44bd2b3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.279070 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-server-conf" (OuterVolumeSpecName: "server-conf") pod "493425e2-b4c7-4a1b-af05-44b4f44bd2b3" (UID: "493425e2-b4c7-4a1b-af05-44b4f44bd2b3"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.305248 4894 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-server-conf\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.305292 4894 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.305303 4894 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-pod-info\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.305314 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.305327 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.305338 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.305350 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.305376 4894 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.305387 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b22pt\" (UniqueName: \"kubernetes.io/projected/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-kube-api-access-b22pt\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.305398 4894 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.333165 4894 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.335179 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "493425e2-b4c7-4a1b-af05-44b4f44bd2b3" (UID: "493425e2-b4c7-4a1b-af05-44b4f44bd2b3"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.396241 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="44fdaebd-a96c-47f9-a421-a0984f78972e" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.406771 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/493425e2-b4c7-4a1b-af05-44b4f44bd2b3-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.406807 4894 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.887414 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"493425e2-b4c7-4a1b-af05-44b4f44bd2b3","Type":"ContainerDied","Data":"072ef85aa8f529936442b0ea2924d94a9451add40a7d18b7bb424f8bee94fb36"} Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.887493 4894 scope.go:117] "RemoveContainer" containerID="521f1e5dfce4a639e4244abf43400cbc86ec270828894f24b2104650d695e1f3" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.887495 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.908152 4894 scope.go:117] "RemoveContainer" containerID="e0cc90efd20a8df80e1313ad14be34aed239c406e39225fc092a39ec293bc945" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.927176 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.933966 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.966845 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 15:53:35 crc kubenswrapper[4894]: E1209 15:53:35.967306 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="493425e2-b4c7-4a1b-af05-44b4f44bd2b3" containerName="setup-container" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.967333 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="493425e2-b4c7-4a1b-af05-44b4f44bd2b3" containerName="setup-container" Dec 09 15:53:35 crc kubenswrapper[4894]: E1209 15:53:35.967352 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="493425e2-b4c7-4a1b-af05-44b4f44bd2b3" containerName="rabbitmq" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.967361 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="493425e2-b4c7-4a1b-af05-44b4f44bd2b3" containerName="rabbitmq" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.967546 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="493425e2-b4c7-4a1b-af05-44b4f44bd2b3" containerName="rabbitmq" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.968614 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.970368 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.970537 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.971723 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.971722 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.971882 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-6dzck" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.971997 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.976630 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Dec 09 15:53:35 crc kubenswrapper[4894]: I1209 15:53:35.983962 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.027427 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/07e75d72-1c2a-4d08-91fb-f96409eee8b3-pod-info\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.027502 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/07e75d72-1c2a-4d08-91fb-f96409eee8b3-server-conf\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.027555 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/07e75d72-1c2a-4d08-91fb-f96409eee8b3-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.027596 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/07e75d72-1c2a-4d08-91fb-f96409eee8b3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.027614 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/07e75d72-1c2a-4d08-91fb-f96409eee8b3-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.027669 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/07e75d72-1c2a-4d08-91fb-f96409eee8b3-config-data\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.027908 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/07e75d72-1c2a-4d08-91fb-f96409eee8b3-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.027969 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qlcsq\" (UniqueName: \"kubernetes.io/projected/07e75d72-1c2a-4d08-91fb-f96409eee8b3-kube-api-access-qlcsq\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.028004 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/07e75d72-1c2a-4d08-91fb-f96409eee8b3-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.028040 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.028070 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/07e75d72-1c2a-4d08-91fb-f96409eee8b3-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.119010 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="493425e2-b4c7-4a1b-af05-44b4f44bd2b3" path="/var/lib/kubelet/pods/493425e2-b4c7-4a1b-af05-44b4f44bd2b3/volumes" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.129755 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/07e75d72-1c2a-4d08-91fb-f96409eee8b3-server-conf\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.129815 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/07e75d72-1c2a-4d08-91fb-f96409eee8b3-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.129853 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/07e75d72-1c2a-4d08-91fb-f96409eee8b3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.129869 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/07e75d72-1c2a-4d08-91fb-f96409eee8b3-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.129926 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/07e75d72-1c2a-4d08-91fb-f96409eee8b3-config-data\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.129963 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/07e75d72-1c2a-4d08-91fb-f96409eee8b3-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.130003 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qlcsq\" (UniqueName: \"kubernetes.io/projected/07e75d72-1c2a-4d08-91fb-f96409eee8b3-kube-api-access-qlcsq\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.130028 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/07e75d72-1c2a-4d08-91fb-f96409eee8b3-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.130052 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.130079 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/07e75d72-1c2a-4d08-91fb-f96409eee8b3-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.130096 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/07e75d72-1c2a-4d08-91fb-f96409eee8b3-pod-info\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.130953 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/07e75d72-1c2a-4d08-91fb-f96409eee8b3-server-conf\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.131011 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/07e75d72-1c2a-4d08-91fb-f96409eee8b3-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.131212 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/07e75d72-1c2a-4d08-91fb-f96409eee8b3-config-data\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.131264 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/07e75d72-1c2a-4d08-91fb-f96409eee8b3-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.131504 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.132026 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/07e75d72-1c2a-4d08-91fb-f96409eee8b3-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.135407 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/07e75d72-1c2a-4d08-91fb-f96409eee8b3-pod-info\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.135443 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/07e75d72-1c2a-4d08-91fb-f96409eee8b3-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.143134 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/07e75d72-1c2a-4d08-91fb-f96409eee8b3-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.143980 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/07e75d72-1c2a-4d08-91fb-f96409eee8b3-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.153991 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qlcsq\" (UniqueName: \"kubernetes.io/projected/07e75d72-1c2a-4d08-91fb-f96409eee8b3-kube-api-access-qlcsq\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.173287 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"rabbitmq-server-0\" (UID: \"07e75d72-1c2a-4d08-91fb-f96409eee8b3\") " pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.298302 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.446104 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.640304 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-plugins\") pod \"44fdaebd-a96c-47f9-a421-a0984f78972e\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.640399 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/44fdaebd-a96c-47f9-a421-a0984f78972e-erlang-cookie-secret\") pod \"44fdaebd-a96c-47f9-a421-a0984f78972e\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.640443 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-confd\") pod \"44fdaebd-a96c-47f9-a421-a0984f78972e\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.640586 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/44fdaebd-a96c-47f9-a421-a0984f78972e-config-data\") pod \"44fdaebd-a96c-47f9-a421-a0984f78972e\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.640722 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/44fdaebd-a96c-47f9-a421-a0984f78972e-pod-info\") pod \"44fdaebd-a96c-47f9-a421-a0984f78972e\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.640848 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-erlang-cookie\") pod \"44fdaebd-a96c-47f9-a421-a0984f78972e\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.640886 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dhgw8\" (UniqueName: \"kubernetes.io/projected/44fdaebd-a96c-47f9-a421-a0984f78972e-kube-api-access-dhgw8\") pod \"44fdaebd-a96c-47f9-a421-a0984f78972e\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.640925 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/44fdaebd-a96c-47f9-a421-a0984f78972e-plugins-conf\") pod \"44fdaebd-a96c-47f9-a421-a0984f78972e\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.640952 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/44fdaebd-a96c-47f9-a421-a0984f78972e-server-conf\") pod \"44fdaebd-a96c-47f9-a421-a0984f78972e\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.640998 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-tls\") pod \"44fdaebd-a96c-47f9-a421-a0984f78972e\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.641027 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"44fdaebd-a96c-47f9-a421-a0984f78972e\" (UID: \"44fdaebd-a96c-47f9-a421-a0984f78972e\") " Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.643947 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "44fdaebd-a96c-47f9-a421-a0984f78972e" (UID: "44fdaebd-a96c-47f9-a421-a0984f78972e"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.644361 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "44fdaebd-a96c-47f9-a421-a0984f78972e" (UID: "44fdaebd-a96c-47f9-a421-a0984f78972e"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.644479 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44fdaebd-a96c-47f9-a421-a0984f78972e-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "44fdaebd-a96c-47f9-a421-a0984f78972e" (UID: "44fdaebd-a96c-47f9-a421-a0984f78972e"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.651668 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "persistence") pod "44fdaebd-a96c-47f9-a421-a0984f78972e" (UID: "44fdaebd-a96c-47f9-a421-a0984f78972e"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.653169 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/44fdaebd-a96c-47f9-a421-a0984f78972e-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "44fdaebd-a96c-47f9-a421-a0984f78972e" (UID: "44fdaebd-a96c-47f9-a421-a0984f78972e"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.666017 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "44fdaebd-a96c-47f9-a421-a0984f78972e" (UID: "44fdaebd-a96c-47f9-a421-a0984f78972e"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.666498 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44fdaebd-a96c-47f9-a421-a0984f78972e-kube-api-access-dhgw8" (OuterVolumeSpecName: "kube-api-access-dhgw8") pod "44fdaebd-a96c-47f9-a421-a0984f78972e" (UID: "44fdaebd-a96c-47f9-a421-a0984f78972e"). InnerVolumeSpecName "kube-api-access-dhgw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.667198 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/44fdaebd-a96c-47f9-a421-a0984f78972e-pod-info" (OuterVolumeSpecName: "pod-info") pod "44fdaebd-a96c-47f9-a421-a0984f78972e" (UID: "44fdaebd-a96c-47f9-a421-a0984f78972e"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.674909 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44fdaebd-a96c-47f9-a421-a0984f78972e-config-data" (OuterVolumeSpecName: "config-data") pod "44fdaebd-a96c-47f9-a421-a0984f78972e" (UID: "44fdaebd-a96c-47f9-a421-a0984f78972e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.712074 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/44fdaebd-a96c-47f9-a421-a0984f78972e-server-conf" (OuterVolumeSpecName: "server-conf") pod "44fdaebd-a96c-47f9-a421-a0984f78972e" (UID: "44fdaebd-a96c-47f9-a421-a0984f78972e"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.744525 4894 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/44fdaebd-a96c-47f9-a421-a0984f78972e-plugins-conf\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.744558 4894 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/44fdaebd-a96c-47f9-a421-a0984f78972e-server-conf\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.744567 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.744593 4894 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.744606 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.744614 4894 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/44fdaebd-a96c-47f9-a421-a0984f78972e-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.744625 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/44fdaebd-a96c-47f9-a421-a0984f78972e-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.744633 4894 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/44fdaebd-a96c-47f9-a421-a0984f78972e-pod-info\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.744673 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.744684 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dhgw8\" (UniqueName: \"kubernetes.io/projected/44fdaebd-a96c-47f9-a421-a0984f78972e-kube-api-access-dhgw8\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.771291 4894 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.775034 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "44fdaebd-a96c-47f9-a421-a0984f78972e" (UID: "44fdaebd-a96c-47f9-a421-a0984f78972e"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.807741 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.848209 4894 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.848268 4894 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/44fdaebd-a96c-47f9-a421-a0984f78972e-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.899022 4894 generic.go:334] "Generic (PLEG): container finished" podID="44fdaebd-a96c-47f9-a421-a0984f78972e" containerID="0ebd4954c8e209f4c21212c42fd48068abec230259a29738503745230bd00e15" exitCode=0 Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.899084 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.899091 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"44fdaebd-a96c-47f9-a421-a0984f78972e","Type":"ContainerDied","Data":"0ebd4954c8e209f4c21212c42fd48068abec230259a29738503745230bd00e15"} Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.899191 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"44fdaebd-a96c-47f9-a421-a0984f78972e","Type":"ContainerDied","Data":"988760f4ec9858251ac39f4e2bf3163cd78c8a1ca08415b2fcdb5c44091f3c13"} Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.899209 4894 scope.go:117] "RemoveContainer" containerID="0ebd4954c8e209f4c21212c42fd48068abec230259a29738503745230bd00e15" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.900111 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"07e75d72-1c2a-4d08-91fb-f96409eee8b3","Type":"ContainerStarted","Data":"cdf00ac53faca1b627b1aeae5567cd6b1617adc942ea15679ae8eb6be02b32be"} Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.931345 4894 scope.go:117] "RemoveContainer" containerID="89642012cdb0862001c73aa935177e0ea6e93168b497fc358d09b0a6cbda4b1a" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.937036 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.956417 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.971719 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 15:53:36 crc kubenswrapper[4894]: E1209 15:53:36.972135 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44fdaebd-a96c-47f9-a421-a0984f78972e" containerName="rabbitmq" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.972148 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="44fdaebd-a96c-47f9-a421-a0984f78972e" containerName="rabbitmq" Dec 09 15:53:36 crc kubenswrapper[4894]: E1209 15:53:36.972162 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44fdaebd-a96c-47f9-a421-a0984f78972e" containerName="setup-container" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.972170 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="44fdaebd-a96c-47f9-a421-a0984f78972e" containerName="setup-container" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.972357 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="44fdaebd-a96c-47f9-a421-a0984f78972e" containerName="rabbitmq" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.973302 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.977184 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.979085 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.979197 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.979244 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.979300 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.979394 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-74ccw" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.979396 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.984384 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.987331 4894 scope.go:117] "RemoveContainer" containerID="0ebd4954c8e209f4c21212c42fd48068abec230259a29738503745230bd00e15" Dec 09 15:53:36 crc kubenswrapper[4894]: E1209 15:53:36.992237 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0ebd4954c8e209f4c21212c42fd48068abec230259a29738503745230bd00e15\": container with ID starting with 0ebd4954c8e209f4c21212c42fd48068abec230259a29738503745230bd00e15 not found: ID does not exist" containerID="0ebd4954c8e209f4c21212c42fd48068abec230259a29738503745230bd00e15" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.992277 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0ebd4954c8e209f4c21212c42fd48068abec230259a29738503745230bd00e15"} err="failed to get container status \"0ebd4954c8e209f4c21212c42fd48068abec230259a29738503745230bd00e15\": rpc error: code = NotFound desc = could not find container \"0ebd4954c8e209f4c21212c42fd48068abec230259a29738503745230bd00e15\": container with ID starting with 0ebd4954c8e209f4c21212c42fd48068abec230259a29738503745230bd00e15 not found: ID does not exist" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.992303 4894 scope.go:117] "RemoveContainer" containerID="89642012cdb0862001c73aa935177e0ea6e93168b497fc358d09b0a6cbda4b1a" Dec 09 15:53:36 crc kubenswrapper[4894]: E1209 15:53:36.992667 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89642012cdb0862001c73aa935177e0ea6e93168b497fc358d09b0a6cbda4b1a\": container with ID starting with 89642012cdb0862001c73aa935177e0ea6e93168b497fc358d09b0a6cbda4b1a not found: ID does not exist" containerID="89642012cdb0862001c73aa935177e0ea6e93168b497fc358d09b0a6cbda4b1a" Dec 09 15:53:36 crc kubenswrapper[4894]: I1209 15:53:36.992723 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89642012cdb0862001c73aa935177e0ea6e93168b497fc358d09b0a6cbda4b1a"} err="failed to get container status \"89642012cdb0862001c73aa935177e0ea6e93168b497fc358d09b0a6cbda4b1a\": rpc error: code = NotFound desc = could not find container \"89642012cdb0862001c73aa935177e0ea6e93168b497fc358d09b0a6cbda4b1a\": container with ID starting with 89642012cdb0862001c73aa935177e0ea6e93168b497fc358d09b0a6cbda4b1a not found: ID does not exist" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.154765 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0ad2400d-f6ca-4f83-a88f-b1bed6287315-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.155224 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0ad2400d-f6ca-4f83-a88f-b1bed6287315-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.155345 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0ad2400d-f6ca-4f83-a88f-b1bed6287315-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.155487 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0ad2400d-f6ca-4f83-a88f-b1bed6287315-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.155591 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.156030 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0ad2400d-f6ca-4f83-a88f-b1bed6287315-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.156148 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0ad2400d-f6ca-4f83-a88f-b1bed6287315-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.156262 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0ad2400d-f6ca-4f83-a88f-b1bed6287315-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.156450 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0ad2400d-f6ca-4f83-a88f-b1bed6287315-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.156549 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0ad2400d-f6ca-4f83-a88f-b1bed6287315-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.156699 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvwgk\" (UniqueName: \"kubernetes.io/projected/0ad2400d-f6ca-4f83-a88f-b1bed6287315-kube-api-access-tvwgk\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.258351 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0ad2400d-f6ca-4f83-a88f-b1bed6287315-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.258436 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0ad2400d-f6ca-4f83-a88f-b1bed6287315-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.258506 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvwgk\" (UniqueName: \"kubernetes.io/projected/0ad2400d-f6ca-4f83-a88f-b1bed6287315-kube-api-access-tvwgk\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.258592 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0ad2400d-f6ca-4f83-a88f-b1bed6287315-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.258675 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0ad2400d-f6ca-4f83-a88f-b1bed6287315-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.258723 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0ad2400d-f6ca-4f83-a88f-b1bed6287315-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.258792 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0ad2400d-f6ca-4f83-a88f-b1bed6287315-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.258840 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.258915 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0ad2400d-f6ca-4f83-a88f-b1bed6287315-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.258952 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0ad2400d-f6ca-4f83-a88f-b1bed6287315-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.258992 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0ad2400d-f6ca-4f83-a88f-b1bed6287315-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.259170 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0ad2400d-f6ca-4f83-a88f-b1bed6287315-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.259368 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.259478 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0ad2400d-f6ca-4f83-a88f-b1bed6287315-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.260414 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0ad2400d-f6ca-4f83-a88f-b1bed6287315-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.260869 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0ad2400d-f6ca-4f83-a88f-b1bed6287315-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.262260 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0ad2400d-f6ca-4f83-a88f-b1bed6287315-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.263712 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0ad2400d-f6ca-4f83-a88f-b1bed6287315-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.263953 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0ad2400d-f6ca-4f83-a88f-b1bed6287315-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.265522 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0ad2400d-f6ca-4f83-a88f-b1bed6287315-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.265528 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0ad2400d-f6ca-4f83-a88f-b1bed6287315-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.282041 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvwgk\" (UniqueName: \"kubernetes.io/projected/0ad2400d-f6ca-4f83-a88f-b1bed6287315-kube-api-access-tvwgk\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.297705 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0ad2400d-f6ca-4f83-a88f-b1bed6287315\") " pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.304591 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:53:37 crc kubenswrapper[4894]: I1209 15:53:37.809111 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Dec 09 15:53:37 crc kubenswrapper[4894]: W1209 15:53:37.910984 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0ad2400d_f6ca_4f83_a88f_b1bed6287315.slice/crio-664f1de60b7f556e54284fafadea2032f81294436ed0c8680005647859740deb WatchSource:0}: Error finding container 664f1de60b7f556e54284fafadea2032f81294436ed0c8680005647859740deb: Status 404 returned error can't find the container with id 664f1de60b7f556e54284fafadea2032f81294436ed0c8680005647859740deb Dec 09 15:53:38 crc kubenswrapper[4894]: I1209 15:53:38.119820 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44fdaebd-a96c-47f9-a421-a0984f78972e" path="/var/lib/kubelet/pods/44fdaebd-a96c-47f9-a421-a0984f78972e/volumes" Dec 09 15:53:38 crc kubenswrapper[4894]: I1209 15:53:38.927752 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0ad2400d-f6ca-4f83-a88f-b1bed6287315","Type":"ContainerStarted","Data":"664f1de60b7f556e54284fafadea2032f81294436ed0c8680005647859740deb"} Dec 09 15:53:39 crc kubenswrapper[4894]: I1209 15:53:39.940585 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"07e75d72-1c2a-4d08-91fb-f96409eee8b3","Type":"ContainerStarted","Data":"d67fc3718edfed4cbb40a6996e3dc7dd93a6618880b109dce4009c8b884f64ba"} Dec 09 15:53:39 crc kubenswrapper[4894]: I1209 15:53:39.943022 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0ad2400d-f6ca-4f83-a88f-b1bed6287315","Type":"ContainerStarted","Data":"a7cc056e67b1f3c2df646883fb318bd467182806cb4db2eb48ca67f11edf8383"} Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.161340 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7bf6f4788c-5kp8h"] Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.164018 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.167098 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-edpm-ipam" Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.179833 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bf6f4788c-5kp8h"] Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.314912 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-dns-svc\") pod \"dnsmasq-dns-7bf6f4788c-5kp8h\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.315024 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-ovsdbserver-nb\") pod \"dnsmasq-dns-7bf6f4788c-5kp8h\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.315051 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xqhl8\" (UniqueName: \"kubernetes.io/projected/31784b7d-7796-47f2-8855-1b2cd8d30cde-kube-api-access-xqhl8\") pod \"dnsmasq-dns-7bf6f4788c-5kp8h\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.315077 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-openstack-edpm-ipam\") pod \"dnsmasq-dns-7bf6f4788c-5kp8h\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.315106 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-ovsdbserver-sb\") pod \"dnsmasq-dns-7bf6f4788c-5kp8h\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.315174 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-config\") pod \"dnsmasq-dns-7bf6f4788c-5kp8h\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.417423 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-dns-svc\") pod \"dnsmasq-dns-7bf6f4788c-5kp8h\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.417678 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-ovsdbserver-nb\") pod \"dnsmasq-dns-7bf6f4788c-5kp8h\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.417750 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xqhl8\" (UniqueName: \"kubernetes.io/projected/31784b7d-7796-47f2-8855-1b2cd8d30cde-kube-api-access-xqhl8\") pod \"dnsmasq-dns-7bf6f4788c-5kp8h\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.417835 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-openstack-edpm-ipam\") pod \"dnsmasq-dns-7bf6f4788c-5kp8h\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.417927 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-ovsdbserver-sb\") pod \"dnsmasq-dns-7bf6f4788c-5kp8h\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.417997 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-config\") pod \"dnsmasq-dns-7bf6f4788c-5kp8h\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.419269 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-dns-svc\") pod \"dnsmasq-dns-7bf6f4788c-5kp8h\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.419360 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-ovsdbserver-sb\") pod \"dnsmasq-dns-7bf6f4788c-5kp8h\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.419614 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-ovsdbserver-nb\") pod \"dnsmasq-dns-7bf6f4788c-5kp8h\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.419750 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-openstack-edpm-ipam\") pod \"dnsmasq-dns-7bf6f4788c-5kp8h\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.420401 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-config\") pod \"dnsmasq-dns-7bf6f4788c-5kp8h\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.460766 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xqhl8\" (UniqueName: \"kubernetes.io/projected/31784b7d-7796-47f2-8855-1b2cd8d30cde-kube-api-access-xqhl8\") pod \"dnsmasq-dns-7bf6f4788c-5kp8h\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:53:40 crc kubenswrapper[4894]: I1209 15:53:40.480837 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:53:41 crc kubenswrapper[4894]: W1209 15:53:40.999779 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31784b7d_7796_47f2_8855_1b2cd8d30cde.slice/crio-51ca8d41cfced2407bbdda427528d649656b4c09dcc7f0bffdc14e2df97ef2db WatchSource:0}: Error finding container 51ca8d41cfced2407bbdda427528d649656b4c09dcc7f0bffdc14e2df97ef2db: Status 404 returned error can't find the container with id 51ca8d41cfced2407bbdda427528d649656b4c09dcc7f0bffdc14e2df97ef2db Dec 09 15:53:41 crc kubenswrapper[4894]: I1209 15:53:40.999800 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7bf6f4788c-5kp8h"] Dec 09 15:53:41 crc kubenswrapper[4894]: I1209 15:53:41.963894 4894 generic.go:334] "Generic (PLEG): container finished" podID="31784b7d-7796-47f2-8855-1b2cd8d30cde" containerID="cfcb5b00bc65a004d86f214ffee1ea7d2ddfa27e16102c5e1590175d52497b21" exitCode=0 Dec 09 15:53:41 crc kubenswrapper[4894]: I1209 15:53:41.963964 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" event={"ID":"31784b7d-7796-47f2-8855-1b2cd8d30cde","Type":"ContainerDied","Data":"cfcb5b00bc65a004d86f214ffee1ea7d2ddfa27e16102c5e1590175d52497b21"} Dec 09 15:53:41 crc kubenswrapper[4894]: I1209 15:53:41.964137 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" event={"ID":"31784b7d-7796-47f2-8855-1b2cd8d30cde","Type":"ContainerStarted","Data":"51ca8d41cfced2407bbdda427528d649656b4c09dcc7f0bffdc14e2df97ef2db"} Dec 09 15:53:42 crc kubenswrapper[4894]: I1209 15:53:42.257775 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:53:42 crc kubenswrapper[4894]: I1209 15:53:42.258154 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:53:42 crc kubenswrapper[4894]: I1209 15:53:42.975957 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" event={"ID":"31784b7d-7796-47f2-8855-1b2cd8d30cde","Type":"ContainerStarted","Data":"be410e851abd29b8d6e9a93b4cb5003aeb4bc19a67ea6964c88c916151437eeb"} Dec 09 15:53:42 crc kubenswrapper[4894]: I1209 15:53:42.976253 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:53:43 crc kubenswrapper[4894]: I1209 15:53:43.002940 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" podStartSLOduration=3.002916377 podStartE2EDuration="3.002916377s" podCreationTimestamp="2025-12-09 15:53:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:53:42.999273748 +0000 UTC m=+1317.318484447" watchObservedRunningTime="2025-12-09 15:53:43.002916377 +0000 UTC m=+1317.322127046" Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.482531 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.562826 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f69c5c76f-kcfmr"] Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.563677 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" podUID="70af2821-3434-450b-b5e1-d876b1ec90ed" containerName="dnsmasq-dns" containerID="cri-o://aa6796049964581c30dd9bf1a036c7e6210221c7c2c7b9c76fdcdcbec36a2055" gracePeriod=10 Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.781265 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5dfc84b95f-nsxl6"] Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.782863 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.794729 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dfc84b95f-nsxl6"] Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.839874 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-openstack-edpm-ipam\") pod \"dnsmasq-dns-5dfc84b95f-nsxl6\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.839944 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-ovsdbserver-sb\") pod \"dnsmasq-dns-5dfc84b95f-nsxl6\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.840001 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-dns-svc\") pod \"dnsmasq-dns-5dfc84b95f-nsxl6\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.840022 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-ovsdbserver-nb\") pod \"dnsmasq-dns-5dfc84b95f-nsxl6\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.840083 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rpjpm\" (UniqueName: \"kubernetes.io/projected/ba44767a-6e2a-422d-84c1-17b60bb6620b-kube-api-access-rpjpm\") pod \"dnsmasq-dns-5dfc84b95f-nsxl6\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.840242 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-config\") pod \"dnsmasq-dns-5dfc84b95f-nsxl6\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.942026 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-config\") pod \"dnsmasq-dns-5dfc84b95f-nsxl6\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.942125 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-openstack-edpm-ipam\") pod \"dnsmasq-dns-5dfc84b95f-nsxl6\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.942177 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-ovsdbserver-sb\") pod \"dnsmasq-dns-5dfc84b95f-nsxl6\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.942240 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-dns-svc\") pod \"dnsmasq-dns-5dfc84b95f-nsxl6\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.942269 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-ovsdbserver-nb\") pod \"dnsmasq-dns-5dfc84b95f-nsxl6\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.943149 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-config\") pod \"dnsmasq-dns-5dfc84b95f-nsxl6\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.943168 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-openstack-edpm-ipam\") pod \"dnsmasq-dns-5dfc84b95f-nsxl6\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.943190 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-dns-svc\") pod \"dnsmasq-dns-5dfc84b95f-nsxl6\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.943357 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-ovsdbserver-nb\") pod \"dnsmasq-dns-5dfc84b95f-nsxl6\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.943458 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-ovsdbserver-sb\") pod \"dnsmasq-dns-5dfc84b95f-nsxl6\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.943532 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rpjpm\" (UniqueName: \"kubernetes.io/projected/ba44767a-6e2a-422d-84c1-17b60bb6620b-kube-api-access-rpjpm\") pod \"dnsmasq-dns-5dfc84b95f-nsxl6\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 15:53:50 crc kubenswrapper[4894]: I1209 15:53:50.961408 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rpjpm\" (UniqueName: \"kubernetes.io/projected/ba44767a-6e2a-422d-84c1-17b60bb6620b-kube-api-access-rpjpm\") pod \"dnsmasq-dns-5dfc84b95f-nsxl6\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 15:53:51 crc kubenswrapper[4894]: I1209 15:53:51.063685 4894 generic.go:334] "Generic (PLEG): container finished" podID="70af2821-3434-450b-b5e1-d876b1ec90ed" containerID="aa6796049964581c30dd9bf1a036c7e6210221c7c2c7b9c76fdcdcbec36a2055" exitCode=0 Dec 09 15:53:51 crc kubenswrapper[4894]: I1209 15:53:51.063739 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" event={"ID":"70af2821-3434-450b-b5e1-d876b1ec90ed","Type":"ContainerDied","Data":"aa6796049964581c30dd9bf1a036c7e6210221c7c2c7b9c76fdcdcbec36a2055"} Dec 09 15:53:51 crc kubenswrapper[4894]: I1209 15:53:51.098229 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 15:53:51 crc kubenswrapper[4894]: I1209 15:53:51.557140 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5dfc84b95f-nsxl6"] Dec 09 15:53:51 crc kubenswrapper[4894]: I1209 15:53:51.579126 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" Dec 09 15:53:51 crc kubenswrapper[4894]: I1209 15:53:51.655502 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-ovsdbserver-sb\") pod \"70af2821-3434-450b-b5e1-d876b1ec90ed\" (UID: \"70af2821-3434-450b-b5e1-d876b1ec90ed\") " Dec 09 15:53:51 crc kubenswrapper[4894]: I1209 15:53:51.655556 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-config\") pod \"70af2821-3434-450b-b5e1-d876b1ec90ed\" (UID: \"70af2821-3434-450b-b5e1-d876b1ec90ed\") " Dec 09 15:53:51 crc kubenswrapper[4894]: I1209 15:53:51.655677 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-dns-svc\") pod \"70af2821-3434-450b-b5e1-d876b1ec90ed\" (UID: \"70af2821-3434-450b-b5e1-d876b1ec90ed\") " Dec 09 15:53:51 crc kubenswrapper[4894]: I1209 15:53:51.655708 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j2rvm\" (UniqueName: \"kubernetes.io/projected/70af2821-3434-450b-b5e1-d876b1ec90ed-kube-api-access-j2rvm\") pod \"70af2821-3434-450b-b5e1-d876b1ec90ed\" (UID: \"70af2821-3434-450b-b5e1-d876b1ec90ed\") " Dec 09 15:53:51 crc kubenswrapper[4894]: I1209 15:53:51.655747 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-ovsdbserver-nb\") pod \"70af2821-3434-450b-b5e1-d876b1ec90ed\" (UID: \"70af2821-3434-450b-b5e1-d876b1ec90ed\") " Dec 09 15:53:51 crc kubenswrapper[4894]: I1209 15:53:51.659413 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/70af2821-3434-450b-b5e1-d876b1ec90ed-kube-api-access-j2rvm" (OuterVolumeSpecName: "kube-api-access-j2rvm") pod "70af2821-3434-450b-b5e1-d876b1ec90ed" (UID: "70af2821-3434-450b-b5e1-d876b1ec90ed"). InnerVolumeSpecName "kube-api-access-j2rvm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:53:51 crc kubenswrapper[4894]: I1209 15:53:51.702280 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "70af2821-3434-450b-b5e1-d876b1ec90ed" (UID: "70af2821-3434-450b-b5e1-d876b1ec90ed"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:53:51 crc kubenswrapper[4894]: I1209 15:53:51.714538 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "70af2821-3434-450b-b5e1-d876b1ec90ed" (UID: "70af2821-3434-450b-b5e1-d876b1ec90ed"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:53:51 crc kubenswrapper[4894]: I1209 15:53:51.721280 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "70af2821-3434-450b-b5e1-d876b1ec90ed" (UID: "70af2821-3434-450b-b5e1-d876b1ec90ed"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:53:51 crc kubenswrapper[4894]: I1209 15:53:51.723620 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-config" (OuterVolumeSpecName: "config") pod "70af2821-3434-450b-b5e1-d876b1ec90ed" (UID: "70af2821-3434-450b-b5e1-d876b1ec90ed"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:53:51 crc kubenswrapper[4894]: I1209 15:53:51.757812 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:51 crc kubenswrapper[4894]: I1209 15:53:51.757845 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j2rvm\" (UniqueName: \"kubernetes.io/projected/70af2821-3434-450b-b5e1-d876b1ec90ed-kube-api-access-j2rvm\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:51 crc kubenswrapper[4894]: I1209 15:53:51.757859 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:51 crc kubenswrapper[4894]: I1209 15:53:51.757868 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:51 crc kubenswrapper[4894]: I1209 15:53:51.757876 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/70af2821-3434-450b-b5e1-d876b1ec90ed-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:53:52 crc kubenswrapper[4894]: I1209 15:53:52.073198 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" Dec 09 15:53:52 crc kubenswrapper[4894]: I1209 15:53:52.073208 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" event={"ID":"70af2821-3434-450b-b5e1-d876b1ec90ed","Type":"ContainerDied","Data":"84b274a4c36e1454582a33d579d1e29d04f3278ec53a4df6dfc782a13d54a1bb"} Dec 09 15:53:52 crc kubenswrapper[4894]: I1209 15:53:52.073844 4894 scope.go:117] "RemoveContainer" containerID="aa6796049964581c30dd9bf1a036c7e6210221c7c2c7b9c76fdcdcbec36a2055" Dec 09 15:53:52 crc kubenswrapper[4894]: I1209 15:53:52.077326 4894 generic.go:334] "Generic (PLEG): container finished" podID="ba44767a-6e2a-422d-84c1-17b60bb6620b" containerID="55f1157d0f458b89f380f3c7ac519777160861ab1b12eff4c262aa4d8f204939" exitCode=0 Dec 09 15:53:52 crc kubenswrapper[4894]: I1209 15:53:52.077586 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" event={"ID":"ba44767a-6e2a-422d-84c1-17b60bb6620b","Type":"ContainerDied","Data":"55f1157d0f458b89f380f3c7ac519777160861ab1b12eff4c262aa4d8f204939"} Dec 09 15:53:52 crc kubenswrapper[4894]: I1209 15:53:52.077612 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" event={"ID":"ba44767a-6e2a-422d-84c1-17b60bb6620b","Type":"ContainerStarted","Data":"1204c89e5cb2d21c4e5aed06c10c2d632add2adb3cc16e446e7b4d8cf64b48b2"} Dec 09 15:53:52 crc kubenswrapper[4894]: I1209 15:53:52.103058 4894 scope.go:117] "RemoveContainer" containerID="5e78c3b981df33475f764aa60405f26aa363f841a5465e8fd5ff95a290ef4fa9" Dec 09 15:53:53 crc kubenswrapper[4894]: I1209 15:53:53.092249 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" event={"ID":"ba44767a-6e2a-422d-84c1-17b60bb6620b","Type":"ContainerStarted","Data":"55c2323af6293d64ad3288f86721689faf620a46518012bedee4f1828e01756f"} Dec 09 15:53:53 crc kubenswrapper[4894]: I1209 15:53:53.093724 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 15:53:53 crc kubenswrapper[4894]: I1209 15:53:53.119017 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" podStartSLOduration=3.118987144 podStartE2EDuration="3.118987144s" podCreationTimestamp="2025-12-09 15:53:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:53:53.113697581 +0000 UTC m=+1327.432908270" watchObservedRunningTime="2025-12-09 15:53:53.118987144 +0000 UTC m=+1327.438197903" Dec 09 15:54:01 crc kubenswrapper[4894]: I1209 15:54:01.100113 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 15:54:01 crc kubenswrapper[4894]: I1209 15:54:01.179608 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bf6f4788c-5kp8h"] Dec 09 15:54:01 crc kubenswrapper[4894]: I1209 15:54:01.179905 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" podUID="31784b7d-7796-47f2-8855-1b2cd8d30cde" containerName="dnsmasq-dns" containerID="cri-o://be410e851abd29b8d6e9a93b4cb5003aeb4bc19a67ea6964c88c916151437eeb" gracePeriod=10 Dec 09 15:54:01 crc kubenswrapper[4894]: I1209 15:54:01.603424 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:54:01 crc kubenswrapper[4894]: I1209 15:54:01.820675 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-ovsdbserver-nb\") pod \"31784b7d-7796-47f2-8855-1b2cd8d30cde\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " Dec 09 15:54:01 crc kubenswrapper[4894]: I1209 15:54:01.820720 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-config\") pod \"31784b7d-7796-47f2-8855-1b2cd8d30cde\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " Dec 09 15:54:01 crc kubenswrapper[4894]: I1209 15:54:01.820754 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-openstack-edpm-ipam\") pod \"31784b7d-7796-47f2-8855-1b2cd8d30cde\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " Dec 09 15:54:01 crc kubenswrapper[4894]: I1209 15:54:01.820804 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-dns-svc\") pod \"31784b7d-7796-47f2-8855-1b2cd8d30cde\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " Dec 09 15:54:01 crc kubenswrapper[4894]: I1209 15:54:01.820836 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-ovsdbserver-sb\") pod \"31784b7d-7796-47f2-8855-1b2cd8d30cde\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " Dec 09 15:54:01 crc kubenswrapper[4894]: I1209 15:54:01.820865 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xqhl8\" (UniqueName: \"kubernetes.io/projected/31784b7d-7796-47f2-8855-1b2cd8d30cde-kube-api-access-xqhl8\") pod \"31784b7d-7796-47f2-8855-1b2cd8d30cde\" (UID: \"31784b7d-7796-47f2-8855-1b2cd8d30cde\") " Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.015367 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31784b7d-7796-47f2-8855-1b2cd8d30cde-kube-api-access-xqhl8" (OuterVolumeSpecName: "kube-api-access-xqhl8") pod "31784b7d-7796-47f2-8855-1b2cd8d30cde" (UID: "31784b7d-7796-47f2-8855-1b2cd8d30cde"). InnerVolumeSpecName "kube-api-access-xqhl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.033670 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xqhl8\" (UniqueName: \"kubernetes.io/projected/31784b7d-7796-47f2-8855-1b2cd8d30cde-kube-api-access-xqhl8\") on node \"crc\" DevicePath \"\"" Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.051832 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-config" (OuterVolumeSpecName: "config") pod "31784b7d-7796-47f2-8855-1b2cd8d30cde" (UID: "31784b7d-7796-47f2-8855-1b2cd8d30cde"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.057047 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "31784b7d-7796-47f2-8855-1b2cd8d30cde" (UID: "31784b7d-7796-47f2-8855-1b2cd8d30cde"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.057496 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "31784b7d-7796-47f2-8855-1b2cd8d30cde" (UID: "31784b7d-7796-47f2-8855-1b2cd8d30cde"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.062218 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "31784b7d-7796-47f2-8855-1b2cd8d30cde" (UID: "31784b7d-7796-47f2-8855-1b2cd8d30cde"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.062265 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "31784b7d-7796-47f2-8855-1b2cd8d30cde" (UID: "31784b7d-7796-47f2-8855-1b2cd8d30cde"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.135520 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.135563 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-config\") on node \"crc\" DevicePath \"\"" Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.135575 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.135587 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.135596 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/31784b7d-7796-47f2-8855-1b2cd8d30cde-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.182830 4894 generic.go:334] "Generic (PLEG): container finished" podID="31784b7d-7796-47f2-8855-1b2cd8d30cde" containerID="be410e851abd29b8d6e9a93b4cb5003aeb4bc19a67ea6964c88c916151437eeb" exitCode=0 Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.182887 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" event={"ID":"31784b7d-7796-47f2-8855-1b2cd8d30cde","Type":"ContainerDied","Data":"be410e851abd29b8d6e9a93b4cb5003aeb4bc19a67ea6964c88c916151437eeb"} Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.182927 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" event={"ID":"31784b7d-7796-47f2-8855-1b2cd8d30cde","Type":"ContainerDied","Data":"51ca8d41cfced2407bbdda427528d649656b4c09dcc7f0bffdc14e2df97ef2db"} Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.182927 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7bf6f4788c-5kp8h" Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.182950 4894 scope.go:117] "RemoveContainer" containerID="be410e851abd29b8d6e9a93b4cb5003aeb4bc19a67ea6964c88c916151437eeb" Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.215856 4894 scope.go:117] "RemoveContainer" containerID="cfcb5b00bc65a004d86f214ffee1ea7d2ddfa27e16102c5e1590175d52497b21" Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.222267 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7bf6f4788c-5kp8h"] Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.238352 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7bf6f4788c-5kp8h"] Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.241107 4894 scope.go:117] "RemoveContainer" containerID="be410e851abd29b8d6e9a93b4cb5003aeb4bc19a67ea6964c88c916151437eeb" Dec 09 15:54:02 crc kubenswrapper[4894]: E1209 15:54:02.241691 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be410e851abd29b8d6e9a93b4cb5003aeb4bc19a67ea6964c88c916151437eeb\": container with ID starting with be410e851abd29b8d6e9a93b4cb5003aeb4bc19a67ea6964c88c916151437eeb not found: ID does not exist" containerID="be410e851abd29b8d6e9a93b4cb5003aeb4bc19a67ea6964c88c916151437eeb" Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.241730 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be410e851abd29b8d6e9a93b4cb5003aeb4bc19a67ea6964c88c916151437eeb"} err="failed to get container status \"be410e851abd29b8d6e9a93b4cb5003aeb4bc19a67ea6964c88c916151437eeb\": rpc error: code = NotFound desc = could not find container \"be410e851abd29b8d6e9a93b4cb5003aeb4bc19a67ea6964c88c916151437eeb\": container with ID starting with be410e851abd29b8d6e9a93b4cb5003aeb4bc19a67ea6964c88c916151437eeb not found: ID does not exist" Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.241757 4894 scope.go:117] "RemoveContainer" containerID="cfcb5b00bc65a004d86f214ffee1ea7d2ddfa27e16102c5e1590175d52497b21" Dec 09 15:54:02 crc kubenswrapper[4894]: E1209 15:54:02.242047 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cfcb5b00bc65a004d86f214ffee1ea7d2ddfa27e16102c5e1590175d52497b21\": container with ID starting with cfcb5b00bc65a004d86f214ffee1ea7d2ddfa27e16102c5e1590175d52497b21 not found: ID does not exist" containerID="cfcb5b00bc65a004d86f214ffee1ea7d2ddfa27e16102c5e1590175d52497b21" Dec 09 15:54:02 crc kubenswrapper[4894]: I1209 15:54:02.242072 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cfcb5b00bc65a004d86f214ffee1ea7d2ddfa27e16102c5e1590175d52497b21"} err="failed to get container status \"cfcb5b00bc65a004d86f214ffee1ea7d2ddfa27e16102c5e1590175d52497b21\": rpc error: code = NotFound desc = could not find container \"cfcb5b00bc65a004d86f214ffee1ea7d2ddfa27e16102c5e1590175d52497b21\": container with ID starting with cfcb5b00bc65a004d86f214ffee1ea7d2ddfa27e16102c5e1590175d52497b21 not found: ID does not exist" Dec 09 15:54:04 crc kubenswrapper[4894]: I1209 15:54:04.116075 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31784b7d-7796-47f2-8855-1b2cd8d30cde" path="/var/lib/kubelet/pods/31784b7d-7796-47f2-8855-1b2cd8d30cde/volumes" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.271219 4894 generic.go:334] "Generic (PLEG): container finished" podID="07e75d72-1c2a-4d08-91fb-f96409eee8b3" containerID="d67fc3718edfed4cbb40a6996e3dc7dd93a6618880b109dce4009c8b884f64ba" exitCode=0 Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.271771 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"07e75d72-1c2a-4d08-91fb-f96409eee8b3","Type":"ContainerDied","Data":"d67fc3718edfed4cbb40a6996e3dc7dd93a6618880b109dce4009c8b884f64ba"} Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.391138 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44"] Dec 09 15:54:11 crc kubenswrapper[4894]: E1209 15:54:11.402920 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31784b7d-7796-47f2-8855-1b2cd8d30cde" containerName="dnsmasq-dns" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.402959 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="31784b7d-7796-47f2-8855-1b2cd8d30cde" containerName="dnsmasq-dns" Dec 09 15:54:11 crc kubenswrapper[4894]: E1209 15:54:11.402992 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31784b7d-7796-47f2-8855-1b2cd8d30cde" containerName="init" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.402999 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="31784b7d-7796-47f2-8855-1b2cd8d30cde" containerName="init" Dec 09 15:54:11 crc kubenswrapper[4894]: E1209 15:54:11.403023 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70af2821-3434-450b-b5e1-d876b1ec90ed" containerName="init" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.403031 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="70af2821-3434-450b-b5e1-d876b1ec90ed" containerName="init" Dec 09 15:54:11 crc kubenswrapper[4894]: E1209 15:54:11.403044 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="70af2821-3434-450b-b5e1-d876b1ec90ed" containerName="dnsmasq-dns" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.403051 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="70af2821-3434-450b-b5e1-d876b1ec90ed" containerName="dnsmasq-dns" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.403278 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="70af2821-3434-450b-b5e1-d876b1ec90ed" containerName="dnsmasq-dns" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.403290 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="31784b7d-7796-47f2-8855-1b2cd8d30cde" containerName="dnsmasq-dns" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.404194 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.411275 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.411703 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.412160 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.412576 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.416453 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44"] Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.517506 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3577fde1-09d2-456c-aaf8-797e1325775c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44\" (UID: \"3577fde1-09d2-456c-aaf8-797e1325775c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.517564 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3577fde1-09d2-456c-aaf8-797e1325775c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44\" (UID: \"3577fde1-09d2-456c-aaf8-797e1325775c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.517730 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3577fde1-09d2-456c-aaf8-797e1325775c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44\" (UID: \"3577fde1-09d2-456c-aaf8-797e1325775c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.517769 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8mhjr\" (UniqueName: \"kubernetes.io/projected/3577fde1-09d2-456c-aaf8-797e1325775c-kube-api-access-8mhjr\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44\" (UID: \"3577fde1-09d2-456c-aaf8-797e1325775c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.619577 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3577fde1-09d2-456c-aaf8-797e1325775c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44\" (UID: \"3577fde1-09d2-456c-aaf8-797e1325775c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.619641 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8mhjr\" (UniqueName: \"kubernetes.io/projected/3577fde1-09d2-456c-aaf8-797e1325775c-kube-api-access-8mhjr\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44\" (UID: \"3577fde1-09d2-456c-aaf8-797e1325775c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.619709 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3577fde1-09d2-456c-aaf8-797e1325775c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44\" (UID: \"3577fde1-09d2-456c-aaf8-797e1325775c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.619740 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3577fde1-09d2-456c-aaf8-797e1325775c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44\" (UID: \"3577fde1-09d2-456c-aaf8-797e1325775c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.624010 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3577fde1-09d2-456c-aaf8-797e1325775c-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44\" (UID: \"3577fde1-09d2-456c-aaf8-797e1325775c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.624165 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3577fde1-09d2-456c-aaf8-797e1325775c-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44\" (UID: \"3577fde1-09d2-456c-aaf8-797e1325775c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.626275 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3577fde1-09d2-456c-aaf8-797e1325775c-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44\" (UID: \"3577fde1-09d2-456c-aaf8-797e1325775c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.641475 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8mhjr\" (UniqueName: \"kubernetes.io/projected/3577fde1-09d2-456c-aaf8-797e1325775c-kube-api-access-8mhjr\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44\" (UID: \"3577fde1-09d2-456c-aaf8-797e1325775c\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44" Dec 09 15:54:11 crc kubenswrapper[4894]: I1209 15:54:11.807420 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44" Dec 09 15:54:12 crc kubenswrapper[4894]: I1209 15:54:12.257443 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:54:12 crc kubenswrapper[4894]: I1209 15:54:12.258081 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:54:12 crc kubenswrapper[4894]: I1209 15:54:12.258129 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 15:54:12 crc kubenswrapper[4894]: I1209 15:54:12.258893 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"cff4b305c5655888c5cc65c71b639fe907b50486a6daafac98221b58e06cb929"} pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 15:54:12 crc kubenswrapper[4894]: I1209 15:54:12.258955 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" containerID="cri-o://cff4b305c5655888c5cc65c71b639fe907b50486a6daafac98221b58e06cb929" gracePeriod=600 Dec 09 15:54:12 crc kubenswrapper[4894]: I1209 15:54:12.295060 4894 generic.go:334] "Generic (PLEG): container finished" podID="0ad2400d-f6ca-4f83-a88f-b1bed6287315" containerID="a7cc056e67b1f3c2df646883fb318bd467182806cb4db2eb48ca67f11edf8383" exitCode=0 Dec 09 15:54:12 crc kubenswrapper[4894]: I1209 15:54:12.295094 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0ad2400d-f6ca-4f83-a88f-b1bed6287315","Type":"ContainerDied","Data":"a7cc056e67b1f3c2df646883fb318bd467182806cb4db2eb48ca67f11edf8383"} Dec 09 15:54:12 crc kubenswrapper[4894]: I1209 15:54:12.301499 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"07e75d72-1c2a-4d08-91fb-f96409eee8b3","Type":"ContainerStarted","Data":"6b554698721f2bd9bf3856f2b3bc87e6132db15731bfa32b01b5593dff47c711"} Dec 09 15:54:12 crc kubenswrapper[4894]: I1209 15:54:12.302049 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Dec 09 15:54:12 crc kubenswrapper[4894]: I1209 15:54:12.353151 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=37.353107479 podStartE2EDuration="37.353107479s" podCreationTimestamp="2025-12-09 15:53:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:54:12.344430095 +0000 UTC m=+1346.663640764" watchObservedRunningTime="2025-12-09 15:54:12.353107479 +0000 UTC m=+1346.672318148" Dec 09 15:54:12 crc kubenswrapper[4894]: I1209 15:54:12.394383 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44"] Dec 09 15:54:13 crc kubenswrapper[4894]: I1209 15:54:13.312572 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44" event={"ID":"3577fde1-09d2-456c-aaf8-797e1325775c","Type":"ContainerStarted","Data":"5d56593722ebbdd61b4c4e519eab058136c9c4d8341682b488512893d2e12b4c"} Dec 09 15:54:13 crc kubenswrapper[4894]: I1209 15:54:13.314886 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0ad2400d-f6ca-4f83-a88f-b1bed6287315","Type":"ContainerStarted","Data":"486525bcae06e38bc3dc569c50d506a447d2730e8af1045dfab1736f586780dd"} Dec 09 15:54:13 crc kubenswrapper[4894]: I1209 15:54:13.315483 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:54:13 crc kubenswrapper[4894]: I1209 15:54:13.317877 4894 generic.go:334] "Generic (PLEG): container finished" podID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerID="cff4b305c5655888c5cc65c71b639fe907b50486a6daafac98221b58e06cb929" exitCode=0 Dec 09 15:54:13 crc kubenswrapper[4894]: I1209 15:54:13.318091 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerDied","Data":"cff4b305c5655888c5cc65c71b639fe907b50486a6daafac98221b58e06cb929"} Dec 09 15:54:13 crc kubenswrapper[4894]: I1209 15:54:13.318285 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerStarted","Data":"5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449"} Dec 09 15:54:13 crc kubenswrapper[4894]: I1209 15:54:13.318338 4894 scope.go:117] "RemoveContainer" containerID="06d076fbf334b50ab934fbbefedcfc6bb4d75c33f4047f53bb661bb7b87d68ea" Dec 09 15:54:13 crc kubenswrapper[4894]: I1209 15:54:13.435523 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.435506417 podStartE2EDuration="37.435506417s" podCreationTimestamp="2025-12-09 15:53:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 15:54:13.412709752 +0000 UTC m=+1347.731920421" watchObservedRunningTime="2025-12-09 15:54:13.435506417 +0000 UTC m=+1347.754717076" Dec 09 15:54:22 crc kubenswrapper[4894]: I1209 15:54:22.204105 4894 pod_container_manager_linux.go:210] "Failed to delete cgroup paths" cgroupName=["kubepods","besteffort","pod70af2821-3434-450b-b5e1-d876b1ec90ed"] err="unable to destroy cgroup paths for cgroup [kubepods besteffort pod70af2821-3434-450b-b5e1-d876b1ec90ed] : Timed out while waiting for systemd to remove kubepods-besteffort-pod70af2821_3434_450b_b5e1_d876b1ec90ed.slice" Dec 09 15:54:22 crc kubenswrapper[4894]: E1209 15:54:22.204792 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to delete cgroup paths for [kubepods besteffort pod70af2821-3434-450b-b5e1-d876b1ec90ed] : unable to destroy cgroup paths for cgroup [kubepods besteffort pod70af2821-3434-450b-b5e1-d876b1ec90ed] : Timed out while waiting for systemd to remove kubepods-besteffort-pod70af2821_3434_450b_b5e1_d876b1ec90ed.slice" pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" podUID="70af2821-3434-450b-b5e1-d876b1ec90ed" Dec 09 15:54:22 crc kubenswrapper[4894]: I1209 15:54:22.409123 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6f69c5c76f-kcfmr" Dec 09 15:54:22 crc kubenswrapper[4894]: I1209 15:54:22.409106 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44" event={"ID":"3577fde1-09d2-456c-aaf8-797e1325775c","Type":"ContainerStarted","Data":"21f39c8734c83c789b43b891077431c05013a3a2112b43c571a86d62a05ee2a2"} Dec 09 15:54:22 crc kubenswrapper[4894]: I1209 15:54:22.436109 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44" podStartSLOduration=1.983220927 podStartE2EDuration="11.436086792s" podCreationTimestamp="2025-12-09 15:54:11 +0000 UTC" firstStartedPulling="2025-12-09 15:54:12.40799245 +0000 UTC m=+1346.727203119" lastFinishedPulling="2025-12-09 15:54:21.860858315 +0000 UTC m=+1356.180068984" observedRunningTime="2025-12-09 15:54:22.425898667 +0000 UTC m=+1356.745109366" watchObservedRunningTime="2025-12-09 15:54:22.436086792 +0000 UTC m=+1356.755297481" Dec 09 15:54:22 crc kubenswrapper[4894]: I1209 15:54:22.506161 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6f69c5c76f-kcfmr"] Dec 09 15:54:22 crc kubenswrapper[4894]: I1209 15:54:22.514279 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6f69c5c76f-kcfmr"] Dec 09 15:54:24 crc kubenswrapper[4894]: I1209 15:54:24.119821 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="70af2821-3434-450b-b5e1-d876b1ec90ed" path="/var/lib/kubelet/pods/70af2821-3434-450b-b5e1-d876b1ec90ed/volumes" Dec 09 15:54:26 crc kubenswrapper[4894]: I1209 15:54:26.302103 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Dec 09 15:54:27 crc kubenswrapper[4894]: I1209 15:54:27.308851 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Dec 09 15:54:34 crc kubenswrapper[4894]: I1209 15:54:34.522036 4894 generic.go:334] "Generic (PLEG): container finished" podID="3577fde1-09d2-456c-aaf8-797e1325775c" containerID="21f39c8734c83c789b43b891077431c05013a3a2112b43c571a86d62a05ee2a2" exitCode=0 Dec 09 15:54:34 crc kubenswrapper[4894]: I1209 15:54:34.522137 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44" event={"ID":"3577fde1-09d2-456c-aaf8-797e1325775c","Type":"ContainerDied","Data":"21f39c8734c83c789b43b891077431c05013a3a2112b43c571a86d62a05ee2a2"} Dec 09 15:54:35 crc kubenswrapper[4894]: I1209 15:54:35.947055 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.033695 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8mhjr\" (UniqueName: \"kubernetes.io/projected/3577fde1-09d2-456c-aaf8-797e1325775c-kube-api-access-8mhjr\") pod \"3577fde1-09d2-456c-aaf8-797e1325775c\" (UID: \"3577fde1-09d2-456c-aaf8-797e1325775c\") " Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.033782 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3577fde1-09d2-456c-aaf8-797e1325775c-ssh-key\") pod \"3577fde1-09d2-456c-aaf8-797e1325775c\" (UID: \"3577fde1-09d2-456c-aaf8-797e1325775c\") " Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.033983 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3577fde1-09d2-456c-aaf8-797e1325775c-inventory\") pod \"3577fde1-09d2-456c-aaf8-797e1325775c\" (UID: \"3577fde1-09d2-456c-aaf8-797e1325775c\") " Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.034010 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3577fde1-09d2-456c-aaf8-797e1325775c-repo-setup-combined-ca-bundle\") pod \"3577fde1-09d2-456c-aaf8-797e1325775c\" (UID: \"3577fde1-09d2-456c-aaf8-797e1325775c\") " Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.048922 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3577fde1-09d2-456c-aaf8-797e1325775c-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "3577fde1-09d2-456c-aaf8-797e1325775c" (UID: "3577fde1-09d2-456c-aaf8-797e1325775c"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.048948 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3577fde1-09d2-456c-aaf8-797e1325775c-kube-api-access-8mhjr" (OuterVolumeSpecName: "kube-api-access-8mhjr") pod "3577fde1-09d2-456c-aaf8-797e1325775c" (UID: "3577fde1-09d2-456c-aaf8-797e1325775c"). InnerVolumeSpecName "kube-api-access-8mhjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.065970 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3577fde1-09d2-456c-aaf8-797e1325775c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3577fde1-09d2-456c-aaf8-797e1325775c" (UID: "3577fde1-09d2-456c-aaf8-797e1325775c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.078392 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3577fde1-09d2-456c-aaf8-797e1325775c-inventory" (OuterVolumeSpecName: "inventory") pod "3577fde1-09d2-456c-aaf8-797e1325775c" (UID: "3577fde1-09d2-456c-aaf8-797e1325775c"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.138922 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3577fde1-09d2-456c-aaf8-797e1325775c-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.138960 4894 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3577fde1-09d2-456c-aaf8-797e1325775c-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.139000 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8mhjr\" (UniqueName: \"kubernetes.io/projected/3577fde1-09d2-456c-aaf8-797e1325775c-kube-api-access-8mhjr\") on node \"crc\" DevicePath \"\"" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.139014 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3577fde1-09d2-456c-aaf8-797e1325775c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.543250 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44" event={"ID":"3577fde1-09d2-456c-aaf8-797e1325775c","Type":"ContainerDied","Data":"5d56593722ebbdd61b4c4e519eab058136c9c4d8341682b488512893d2e12b4c"} Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.543308 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d56593722ebbdd61b4c4e519eab058136c9c4d8341682b488512893d2e12b4c" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.543322 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.612900 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh"] Dec 09 15:54:36 crc kubenswrapper[4894]: E1209 15:54:36.613247 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3577fde1-09d2-456c-aaf8-797e1325775c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.613261 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="3577fde1-09d2-456c-aaf8-797e1325775c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.613429 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="3577fde1-09d2-456c-aaf8-797e1325775c" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.613964 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.618247 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.618586 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.618673 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.620365 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.666219 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh"] Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.749281 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be481ff5-0866-4a78-84a5-2963c405dd12-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh\" (UID: \"be481ff5-0866-4a78-84a5-2963c405dd12\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.749361 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kqrn8\" (UniqueName: \"kubernetes.io/projected/be481ff5-0866-4a78-84a5-2963c405dd12-kube-api-access-kqrn8\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh\" (UID: \"be481ff5-0866-4a78-84a5-2963c405dd12\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.749447 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be481ff5-0866-4a78-84a5-2963c405dd12-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh\" (UID: \"be481ff5-0866-4a78-84a5-2963c405dd12\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.749501 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be481ff5-0866-4a78-84a5-2963c405dd12-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh\" (UID: \"be481ff5-0866-4a78-84a5-2963c405dd12\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.850742 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be481ff5-0866-4a78-84a5-2963c405dd12-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh\" (UID: \"be481ff5-0866-4a78-84a5-2963c405dd12\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.851030 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be481ff5-0866-4a78-84a5-2963c405dd12-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh\" (UID: \"be481ff5-0866-4a78-84a5-2963c405dd12\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.851269 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be481ff5-0866-4a78-84a5-2963c405dd12-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh\" (UID: \"be481ff5-0866-4a78-84a5-2963c405dd12\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.851419 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kqrn8\" (UniqueName: \"kubernetes.io/projected/be481ff5-0866-4a78-84a5-2963c405dd12-kube-api-access-kqrn8\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh\" (UID: \"be481ff5-0866-4a78-84a5-2963c405dd12\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.857825 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be481ff5-0866-4a78-84a5-2963c405dd12-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh\" (UID: \"be481ff5-0866-4a78-84a5-2963c405dd12\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.861187 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be481ff5-0866-4a78-84a5-2963c405dd12-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh\" (UID: \"be481ff5-0866-4a78-84a5-2963c405dd12\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.864005 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be481ff5-0866-4a78-84a5-2963c405dd12-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh\" (UID: \"be481ff5-0866-4a78-84a5-2963c405dd12\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.866700 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kqrn8\" (UniqueName: \"kubernetes.io/projected/be481ff5-0866-4a78-84a5-2963c405dd12-kube-api-access-kqrn8\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh\" (UID: \"be481ff5-0866-4a78-84a5-2963c405dd12\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh" Dec 09 15:54:36 crc kubenswrapper[4894]: I1209 15:54:36.942279 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh" Dec 09 15:54:37 crc kubenswrapper[4894]: I1209 15:54:37.468481 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh"] Dec 09 15:54:37 crc kubenswrapper[4894]: I1209 15:54:37.554221 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh" event={"ID":"be481ff5-0866-4a78-84a5-2963c405dd12","Type":"ContainerStarted","Data":"5bfa6afb217c5595218b1510eba3d255bc18d7e59ae4e1e77ee602a9f11245a2"} Dec 09 15:54:38 crc kubenswrapper[4894]: I1209 15:54:38.575188 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh" event={"ID":"be481ff5-0866-4a78-84a5-2963c405dd12","Type":"ContainerStarted","Data":"6df01dbe7ee500afeee4fdf4e66c4836d9a37ef755c3d2a40fc7b6646637f264"} Dec 09 15:54:54 crc kubenswrapper[4894]: I1209 15:54:54.304610 4894 scope.go:117] "RemoveContainer" containerID="7031c978f263c0eca3e326f932796ae71acdc505f759c19d24b9ce6a759aa6fb" Dec 09 15:54:54 crc kubenswrapper[4894]: I1209 15:54:54.340695 4894 scope.go:117] "RemoveContainer" containerID="0b513bac6350ca01ee0afa6381af4c9d11918e749aac91ba1bed011460bb7b59" Dec 09 15:54:54 crc kubenswrapper[4894]: I1209 15:54:54.389930 4894 scope.go:117] "RemoveContainer" containerID="d64cbbd062255979137593fdda62ec306b7b9a73e617f4bf24a5728dd1d475c3" Dec 09 15:54:54 crc kubenswrapper[4894]: I1209 15:54:54.432351 4894 scope.go:117] "RemoveContainer" containerID="8127de34e5e54c1b1c5977a3a89313547d28de1afb989a55309d38e2e6488467" Dec 09 15:55:54 crc kubenswrapper[4894]: I1209 15:55:54.552028 4894 scope.go:117] "RemoveContainer" containerID="466350b715f0a001b96f8b05271b4b5a567f93e5f14fc6dd76ac361dc90e20d6" Dec 09 15:56:12 crc kubenswrapper[4894]: I1209 15:56:12.258137 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:56:12 crc kubenswrapper[4894]: I1209 15:56:12.258701 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:56:25 crc kubenswrapper[4894]: I1209 15:56:25.830481 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh" podStartSLOduration=109.429562115 podStartE2EDuration="1m49.830465479s" podCreationTimestamp="2025-12-09 15:54:36 +0000 UTC" firstStartedPulling="2025-12-09 15:54:37.467568863 +0000 UTC m=+1371.786779542" lastFinishedPulling="2025-12-09 15:54:37.868472247 +0000 UTC m=+1372.187682906" observedRunningTime="2025-12-09 15:54:38.597255024 +0000 UTC m=+1372.916465703" watchObservedRunningTime="2025-12-09 15:56:25.830465479 +0000 UTC m=+1480.149676148" Dec 09 15:56:25 crc kubenswrapper[4894]: I1209 15:56:25.839309 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-xvh8p"] Dec 09 15:56:25 crc kubenswrapper[4894]: I1209 15:56:25.841255 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xvh8p" Dec 09 15:56:25 crc kubenswrapper[4894]: I1209 15:56:25.919155 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xvh8p"] Dec 09 15:56:26 crc kubenswrapper[4894]: I1209 15:56:26.019410 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e266cde-2920-4266-9ffa-89684208e5db-utilities\") pod \"redhat-operators-xvh8p\" (UID: \"1e266cde-2920-4266-9ffa-89684208e5db\") " pod="openshift-marketplace/redhat-operators-xvh8p" Dec 09 15:56:26 crc kubenswrapper[4894]: I1209 15:56:26.019549 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b97zt\" (UniqueName: \"kubernetes.io/projected/1e266cde-2920-4266-9ffa-89684208e5db-kube-api-access-b97zt\") pod \"redhat-operators-xvh8p\" (UID: \"1e266cde-2920-4266-9ffa-89684208e5db\") " pod="openshift-marketplace/redhat-operators-xvh8p" Dec 09 15:56:26 crc kubenswrapper[4894]: I1209 15:56:26.019685 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e266cde-2920-4266-9ffa-89684208e5db-catalog-content\") pod \"redhat-operators-xvh8p\" (UID: \"1e266cde-2920-4266-9ffa-89684208e5db\") " pod="openshift-marketplace/redhat-operators-xvh8p" Dec 09 15:56:26 crc kubenswrapper[4894]: I1209 15:56:26.121084 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e266cde-2920-4266-9ffa-89684208e5db-utilities\") pod \"redhat-operators-xvh8p\" (UID: \"1e266cde-2920-4266-9ffa-89684208e5db\") " pod="openshift-marketplace/redhat-operators-xvh8p" Dec 09 15:56:26 crc kubenswrapper[4894]: I1209 15:56:26.121169 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b97zt\" (UniqueName: \"kubernetes.io/projected/1e266cde-2920-4266-9ffa-89684208e5db-kube-api-access-b97zt\") pod \"redhat-operators-xvh8p\" (UID: \"1e266cde-2920-4266-9ffa-89684208e5db\") " pod="openshift-marketplace/redhat-operators-xvh8p" Dec 09 15:56:26 crc kubenswrapper[4894]: I1209 15:56:26.121217 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e266cde-2920-4266-9ffa-89684208e5db-catalog-content\") pod \"redhat-operators-xvh8p\" (UID: \"1e266cde-2920-4266-9ffa-89684208e5db\") " pod="openshift-marketplace/redhat-operators-xvh8p" Dec 09 15:56:26 crc kubenswrapper[4894]: I1209 15:56:26.121544 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e266cde-2920-4266-9ffa-89684208e5db-utilities\") pod \"redhat-operators-xvh8p\" (UID: \"1e266cde-2920-4266-9ffa-89684208e5db\") " pod="openshift-marketplace/redhat-operators-xvh8p" Dec 09 15:56:26 crc kubenswrapper[4894]: I1209 15:56:26.121713 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e266cde-2920-4266-9ffa-89684208e5db-catalog-content\") pod \"redhat-operators-xvh8p\" (UID: \"1e266cde-2920-4266-9ffa-89684208e5db\") " pod="openshift-marketplace/redhat-operators-xvh8p" Dec 09 15:56:26 crc kubenswrapper[4894]: I1209 15:56:26.149551 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b97zt\" (UniqueName: \"kubernetes.io/projected/1e266cde-2920-4266-9ffa-89684208e5db-kube-api-access-b97zt\") pod \"redhat-operators-xvh8p\" (UID: \"1e266cde-2920-4266-9ffa-89684208e5db\") " pod="openshift-marketplace/redhat-operators-xvh8p" Dec 09 15:56:26 crc kubenswrapper[4894]: I1209 15:56:26.167102 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xvh8p" Dec 09 15:56:26 crc kubenswrapper[4894]: I1209 15:56:26.656389 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-xvh8p"] Dec 09 15:56:27 crc kubenswrapper[4894]: I1209 15:56:27.009591 4894 generic.go:334] "Generic (PLEG): container finished" podID="1e266cde-2920-4266-9ffa-89684208e5db" containerID="a1a428676f3de554e52b6886940c47c5f7520950f1f9cb0c5e1dec1593d3f151" exitCode=0 Dec 09 15:56:27 crc kubenswrapper[4894]: I1209 15:56:27.009653 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xvh8p" event={"ID":"1e266cde-2920-4266-9ffa-89684208e5db","Type":"ContainerDied","Data":"a1a428676f3de554e52b6886940c47c5f7520950f1f9cb0c5e1dec1593d3f151"} Dec 09 15:56:27 crc kubenswrapper[4894]: I1209 15:56:27.009682 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xvh8p" event={"ID":"1e266cde-2920-4266-9ffa-89684208e5db","Type":"ContainerStarted","Data":"76516d44e9d3fad9ae73e6adc6b9d6b2348e01d3d8ed0ea9390eacb7ca63612c"} Dec 09 15:56:28 crc kubenswrapper[4894]: I1209 15:56:28.020394 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xvh8p" event={"ID":"1e266cde-2920-4266-9ffa-89684208e5db","Type":"ContainerStarted","Data":"f3af169e41b60a15643abcc5ea97f345719b698a9dc8923f5c090c1142cdec1e"} Dec 09 15:56:30 crc kubenswrapper[4894]: I1209 15:56:30.046567 4894 generic.go:334] "Generic (PLEG): container finished" podID="1e266cde-2920-4266-9ffa-89684208e5db" containerID="f3af169e41b60a15643abcc5ea97f345719b698a9dc8923f5c090c1142cdec1e" exitCode=0 Dec 09 15:56:30 crc kubenswrapper[4894]: I1209 15:56:30.046681 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xvh8p" event={"ID":"1e266cde-2920-4266-9ffa-89684208e5db","Type":"ContainerDied","Data":"f3af169e41b60a15643abcc5ea97f345719b698a9dc8923f5c090c1142cdec1e"} Dec 09 15:56:31 crc kubenswrapper[4894]: I1209 15:56:31.059228 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xvh8p" event={"ID":"1e266cde-2920-4266-9ffa-89684208e5db","Type":"ContainerStarted","Data":"4bcbf5671b3bfba7ee4a8188fffb7f226fb6a3d85cf0fcce68ab06d85d597b98"} Dec 09 15:56:31 crc kubenswrapper[4894]: I1209 15:56:31.086966 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-xvh8p" podStartSLOduration=2.2772990379999998 podStartE2EDuration="6.086947586s" podCreationTimestamp="2025-12-09 15:56:25 +0000 UTC" firstStartedPulling="2025-12-09 15:56:27.011601562 +0000 UTC m=+1481.330812231" lastFinishedPulling="2025-12-09 15:56:30.82125011 +0000 UTC m=+1485.140460779" observedRunningTime="2025-12-09 15:56:31.078765943 +0000 UTC m=+1485.397976632" watchObservedRunningTime="2025-12-09 15:56:31.086947586 +0000 UTC m=+1485.406158255" Dec 09 15:56:36 crc kubenswrapper[4894]: I1209 15:56:36.167964 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-xvh8p" Dec 09 15:56:36 crc kubenswrapper[4894]: I1209 15:56:36.169373 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-xvh8p" Dec 09 15:56:36 crc kubenswrapper[4894]: I1209 15:56:36.216133 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-xvh8p" Dec 09 15:56:37 crc kubenswrapper[4894]: I1209 15:56:37.179040 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-xvh8p" Dec 09 15:56:37 crc kubenswrapper[4894]: I1209 15:56:37.220801 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xvh8p"] Dec 09 15:56:39 crc kubenswrapper[4894]: I1209 15:56:39.130917 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-xvh8p" podUID="1e266cde-2920-4266-9ffa-89684208e5db" containerName="registry-server" containerID="cri-o://4bcbf5671b3bfba7ee4a8188fffb7f226fb6a3d85cf0fcce68ab06d85d597b98" gracePeriod=2 Dec 09 15:56:39 crc kubenswrapper[4894]: I1209 15:56:39.550032 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xvh8p" Dec 09 15:56:39 crc kubenswrapper[4894]: I1209 15:56:39.676414 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e266cde-2920-4266-9ffa-89684208e5db-catalog-content\") pod \"1e266cde-2920-4266-9ffa-89684208e5db\" (UID: \"1e266cde-2920-4266-9ffa-89684208e5db\") " Dec 09 15:56:39 crc kubenswrapper[4894]: I1209 15:56:39.676478 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e266cde-2920-4266-9ffa-89684208e5db-utilities\") pod \"1e266cde-2920-4266-9ffa-89684208e5db\" (UID: \"1e266cde-2920-4266-9ffa-89684208e5db\") " Dec 09 15:56:39 crc kubenswrapper[4894]: I1209 15:56:39.676744 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b97zt\" (UniqueName: \"kubernetes.io/projected/1e266cde-2920-4266-9ffa-89684208e5db-kube-api-access-b97zt\") pod \"1e266cde-2920-4266-9ffa-89684208e5db\" (UID: \"1e266cde-2920-4266-9ffa-89684208e5db\") " Dec 09 15:56:39 crc kubenswrapper[4894]: I1209 15:56:39.677573 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e266cde-2920-4266-9ffa-89684208e5db-utilities" (OuterVolumeSpecName: "utilities") pod "1e266cde-2920-4266-9ffa-89684208e5db" (UID: "1e266cde-2920-4266-9ffa-89684208e5db"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:56:39 crc kubenswrapper[4894]: I1209 15:56:39.685066 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1e266cde-2920-4266-9ffa-89684208e5db-kube-api-access-b97zt" (OuterVolumeSpecName: "kube-api-access-b97zt") pod "1e266cde-2920-4266-9ffa-89684208e5db" (UID: "1e266cde-2920-4266-9ffa-89684208e5db"). InnerVolumeSpecName "kube-api-access-b97zt". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:56:39 crc kubenswrapper[4894]: I1209 15:56:39.778888 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1e266cde-2920-4266-9ffa-89684208e5db-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 15:56:39 crc kubenswrapper[4894]: I1209 15:56:39.778943 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b97zt\" (UniqueName: \"kubernetes.io/projected/1e266cde-2920-4266-9ffa-89684208e5db-kube-api-access-b97zt\") on node \"crc\" DevicePath \"\"" Dec 09 15:56:39 crc kubenswrapper[4894]: I1209 15:56:39.782481 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1e266cde-2920-4266-9ffa-89684208e5db-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1e266cde-2920-4266-9ffa-89684208e5db" (UID: "1e266cde-2920-4266-9ffa-89684208e5db"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:56:39 crc kubenswrapper[4894]: I1209 15:56:39.880315 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1e266cde-2920-4266-9ffa-89684208e5db-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 15:56:40 crc kubenswrapper[4894]: I1209 15:56:40.140858 4894 generic.go:334] "Generic (PLEG): container finished" podID="1e266cde-2920-4266-9ffa-89684208e5db" containerID="4bcbf5671b3bfba7ee4a8188fffb7f226fb6a3d85cf0fcce68ab06d85d597b98" exitCode=0 Dec 09 15:56:40 crc kubenswrapper[4894]: I1209 15:56:40.140904 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xvh8p" event={"ID":"1e266cde-2920-4266-9ffa-89684208e5db","Type":"ContainerDied","Data":"4bcbf5671b3bfba7ee4a8188fffb7f226fb6a3d85cf0fcce68ab06d85d597b98"} Dec 09 15:56:40 crc kubenswrapper[4894]: I1209 15:56:40.140931 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-xvh8p" event={"ID":"1e266cde-2920-4266-9ffa-89684208e5db","Type":"ContainerDied","Data":"76516d44e9d3fad9ae73e6adc6b9d6b2348e01d3d8ed0ea9390eacb7ca63612c"} Dec 09 15:56:40 crc kubenswrapper[4894]: I1209 15:56:40.140949 4894 scope.go:117] "RemoveContainer" containerID="4bcbf5671b3bfba7ee4a8188fffb7f226fb6a3d85cf0fcce68ab06d85d597b98" Dec 09 15:56:40 crc kubenswrapper[4894]: I1209 15:56:40.140981 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-xvh8p" Dec 09 15:56:40 crc kubenswrapper[4894]: I1209 15:56:40.163494 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-xvh8p"] Dec 09 15:56:40 crc kubenswrapper[4894]: I1209 15:56:40.166777 4894 scope.go:117] "RemoveContainer" containerID="f3af169e41b60a15643abcc5ea97f345719b698a9dc8923f5c090c1142cdec1e" Dec 09 15:56:40 crc kubenswrapper[4894]: I1209 15:56:40.170865 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-xvh8p"] Dec 09 15:56:40 crc kubenswrapper[4894]: I1209 15:56:40.199189 4894 scope.go:117] "RemoveContainer" containerID="a1a428676f3de554e52b6886940c47c5f7520950f1f9cb0c5e1dec1593d3f151" Dec 09 15:56:40 crc kubenswrapper[4894]: I1209 15:56:40.226393 4894 scope.go:117] "RemoveContainer" containerID="4bcbf5671b3bfba7ee4a8188fffb7f226fb6a3d85cf0fcce68ab06d85d597b98" Dec 09 15:56:40 crc kubenswrapper[4894]: E1209 15:56:40.227093 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4bcbf5671b3bfba7ee4a8188fffb7f226fb6a3d85cf0fcce68ab06d85d597b98\": container with ID starting with 4bcbf5671b3bfba7ee4a8188fffb7f226fb6a3d85cf0fcce68ab06d85d597b98 not found: ID does not exist" containerID="4bcbf5671b3bfba7ee4a8188fffb7f226fb6a3d85cf0fcce68ab06d85d597b98" Dec 09 15:56:40 crc kubenswrapper[4894]: I1209 15:56:40.227128 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4bcbf5671b3bfba7ee4a8188fffb7f226fb6a3d85cf0fcce68ab06d85d597b98"} err="failed to get container status \"4bcbf5671b3bfba7ee4a8188fffb7f226fb6a3d85cf0fcce68ab06d85d597b98\": rpc error: code = NotFound desc = could not find container \"4bcbf5671b3bfba7ee4a8188fffb7f226fb6a3d85cf0fcce68ab06d85d597b98\": container with ID starting with 4bcbf5671b3bfba7ee4a8188fffb7f226fb6a3d85cf0fcce68ab06d85d597b98 not found: ID does not exist" Dec 09 15:56:40 crc kubenswrapper[4894]: I1209 15:56:40.227155 4894 scope.go:117] "RemoveContainer" containerID="f3af169e41b60a15643abcc5ea97f345719b698a9dc8923f5c090c1142cdec1e" Dec 09 15:56:40 crc kubenswrapper[4894]: E1209 15:56:40.227482 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f3af169e41b60a15643abcc5ea97f345719b698a9dc8923f5c090c1142cdec1e\": container with ID starting with f3af169e41b60a15643abcc5ea97f345719b698a9dc8923f5c090c1142cdec1e not found: ID does not exist" containerID="f3af169e41b60a15643abcc5ea97f345719b698a9dc8923f5c090c1142cdec1e" Dec 09 15:56:40 crc kubenswrapper[4894]: I1209 15:56:40.227523 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f3af169e41b60a15643abcc5ea97f345719b698a9dc8923f5c090c1142cdec1e"} err="failed to get container status \"f3af169e41b60a15643abcc5ea97f345719b698a9dc8923f5c090c1142cdec1e\": rpc error: code = NotFound desc = could not find container \"f3af169e41b60a15643abcc5ea97f345719b698a9dc8923f5c090c1142cdec1e\": container with ID starting with f3af169e41b60a15643abcc5ea97f345719b698a9dc8923f5c090c1142cdec1e not found: ID does not exist" Dec 09 15:56:40 crc kubenswrapper[4894]: I1209 15:56:40.227550 4894 scope.go:117] "RemoveContainer" containerID="a1a428676f3de554e52b6886940c47c5f7520950f1f9cb0c5e1dec1593d3f151" Dec 09 15:56:40 crc kubenswrapper[4894]: E1209 15:56:40.227880 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1a428676f3de554e52b6886940c47c5f7520950f1f9cb0c5e1dec1593d3f151\": container with ID starting with a1a428676f3de554e52b6886940c47c5f7520950f1f9cb0c5e1dec1593d3f151 not found: ID does not exist" containerID="a1a428676f3de554e52b6886940c47c5f7520950f1f9cb0c5e1dec1593d3f151" Dec 09 15:56:40 crc kubenswrapper[4894]: I1209 15:56:40.227911 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1a428676f3de554e52b6886940c47c5f7520950f1f9cb0c5e1dec1593d3f151"} err="failed to get container status \"a1a428676f3de554e52b6886940c47c5f7520950f1f9cb0c5e1dec1593d3f151\": rpc error: code = NotFound desc = could not find container \"a1a428676f3de554e52b6886940c47c5f7520950f1f9cb0c5e1dec1593d3f151\": container with ID starting with a1a428676f3de554e52b6886940c47c5f7520950f1f9cb0c5e1dec1593d3f151 not found: ID does not exist" Dec 09 15:56:42 crc kubenswrapper[4894]: I1209 15:56:42.117431 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1e266cde-2920-4266-9ffa-89684208e5db" path="/var/lib/kubelet/pods/1e266cde-2920-4266-9ffa-89684208e5db/volumes" Dec 09 15:56:42 crc kubenswrapper[4894]: I1209 15:56:42.257266 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:56:42 crc kubenswrapper[4894]: I1209 15:56:42.257775 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:57:12 crc kubenswrapper[4894]: I1209 15:57:12.258044 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 15:57:12 crc kubenswrapper[4894]: I1209 15:57:12.258762 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 15:57:12 crc kubenswrapper[4894]: I1209 15:57:12.258809 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 15:57:12 crc kubenswrapper[4894]: I1209 15:57:12.259627 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449"} pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 15:57:12 crc kubenswrapper[4894]: I1209 15:57:12.259735 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" containerID="cri-o://5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" gracePeriod=600 Dec 09 15:57:12 crc kubenswrapper[4894]: E1209 15:57:12.397478 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 15:57:12 crc kubenswrapper[4894]: I1209 15:57:12.422572 4894 generic.go:334] "Generic (PLEG): container finished" podID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" exitCode=0 Dec 09 15:57:12 crc kubenswrapper[4894]: I1209 15:57:12.422614 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerDied","Data":"5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449"} Dec 09 15:57:12 crc kubenswrapper[4894]: I1209 15:57:12.422664 4894 scope.go:117] "RemoveContainer" containerID="cff4b305c5655888c5cc65c71b639fe907b50486a6daafac98221b58e06cb929" Dec 09 15:57:12 crc kubenswrapper[4894]: I1209 15:57:12.423211 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 15:57:12 crc kubenswrapper[4894]: E1209 15:57:12.423424 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 15:57:22 crc kubenswrapper[4894]: I1209 15:57:22.142692 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-swzkf"] Dec 09 15:57:22 crc kubenswrapper[4894]: E1209 15:57:22.145543 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e266cde-2920-4266-9ffa-89684208e5db" containerName="extract-utilities" Dec 09 15:57:22 crc kubenswrapper[4894]: I1209 15:57:22.145716 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e266cde-2920-4266-9ffa-89684208e5db" containerName="extract-utilities" Dec 09 15:57:22 crc kubenswrapper[4894]: E1209 15:57:22.145862 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e266cde-2920-4266-9ffa-89684208e5db" containerName="registry-server" Dec 09 15:57:22 crc kubenswrapper[4894]: I1209 15:57:22.145975 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e266cde-2920-4266-9ffa-89684208e5db" containerName="registry-server" Dec 09 15:57:22 crc kubenswrapper[4894]: E1209 15:57:22.146074 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1e266cde-2920-4266-9ffa-89684208e5db" containerName="extract-content" Dec 09 15:57:22 crc kubenswrapper[4894]: I1209 15:57:22.146142 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="1e266cde-2920-4266-9ffa-89684208e5db" containerName="extract-content" Dec 09 15:57:22 crc kubenswrapper[4894]: I1209 15:57:22.146490 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="1e266cde-2920-4266-9ffa-89684208e5db" containerName="registry-server" Dec 09 15:57:22 crc kubenswrapper[4894]: I1209 15:57:22.148406 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swzkf" Dec 09 15:57:22 crc kubenswrapper[4894]: I1209 15:57:22.163721 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-swzkf"] Dec 09 15:57:22 crc kubenswrapper[4894]: I1209 15:57:22.182046 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwttm\" (UniqueName: \"kubernetes.io/projected/e8d5cd4c-3ff6-4a99-84c9-f051224f09cd-kube-api-access-fwttm\") pod \"community-operators-swzkf\" (UID: \"e8d5cd4c-3ff6-4a99-84c9-f051224f09cd\") " pod="openshift-marketplace/community-operators-swzkf" Dec 09 15:57:22 crc kubenswrapper[4894]: I1209 15:57:22.182102 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8d5cd4c-3ff6-4a99-84c9-f051224f09cd-utilities\") pod \"community-operators-swzkf\" (UID: \"e8d5cd4c-3ff6-4a99-84c9-f051224f09cd\") " pod="openshift-marketplace/community-operators-swzkf" Dec 09 15:57:22 crc kubenswrapper[4894]: I1209 15:57:22.182176 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8d5cd4c-3ff6-4a99-84c9-f051224f09cd-catalog-content\") pod \"community-operators-swzkf\" (UID: \"e8d5cd4c-3ff6-4a99-84c9-f051224f09cd\") " pod="openshift-marketplace/community-operators-swzkf" Dec 09 15:57:22 crc kubenswrapper[4894]: I1209 15:57:22.284363 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwttm\" (UniqueName: \"kubernetes.io/projected/e8d5cd4c-3ff6-4a99-84c9-f051224f09cd-kube-api-access-fwttm\") pod \"community-operators-swzkf\" (UID: \"e8d5cd4c-3ff6-4a99-84c9-f051224f09cd\") " pod="openshift-marketplace/community-operators-swzkf" Dec 09 15:57:22 crc kubenswrapper[4894]: I1209 15:57:22.284674 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8d5cd4c-3ff6-4a99-84c9-f051224f09cd-utilities\") pod \"community-operators-swzkf\" (UID: \"e8d5cd4c-3ff6-4a99-84c9-f051224f09cd\") " pod="openshift-marketplace/community-operators-swzkf" Dec 09 15:57:22 crc kubenswrapper[4894]: I1209 15:57:22.284785 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8d5cd4c-3ff6-4a99-84c9-f051224f09cd-catalog-content\") pod \"community-operators-swzkf\" (UID: \"e8d5cd4c-3ff6-4a99-84c9-f051224f09cd\") " pod="openshift-marketplace/community-operators-swzkf" Dec 09 15:57:22 crc kubenswrapper[4894]: I1209 15:57:22.287405 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8d5cd4c-3ff6-4a99-84c9-f051224f09cd-catalog-content\") pod \"community-operators-swzkf\" (UID: \"e8d5cd4c-3ff6-4a99-84c9-f051224f09cd\") " pod="openshift-marketplace/community-operators-swzkf" Dec 09 15:57:22 crc kubenswrapper[4894]: I1209 15:57:22.304078 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8d5cd4c-3ff6-4a99-84c9-f051224f09cd-utilities\") pod \"community-operators-swzkf\" (UID: \"e8d5cd4c-3ff6-4a99-84c9-f051224f09cd\") " pod="openshift-marketplace/community-operators-swzkf" Dec 09 15:57:22 crc kubenswrapper[4894]: I1209 15:57:22.342340 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwttm\" (UniqueName: \"kubernetes.io/projected/e8d5cd4c-3ff6-4a99-84c9-f051224f09cd-kube-api-access-fwttm\") pod \"community-operators-swzkf\" (UID: \"e8d5cd4c-3ff6-4a99-84c9-f051224f09cd\") " pod="openshift-marketplace/community-operators-swzkf" Dec 09 15:57:22 crc kubenswrapper[4894]: I1209 15:57:22.504280 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swzkf" Dec 09 15:57:22 crc kubenswrapper[4894]: I1209 15:57:22.992404 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-swzkf"] Dec 09 15:57:23 crc kubenswrapper[4894]: I1209 15:57:23.106246 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 15:57:23 crc kubenswrapper[4894]: E1209 15:57:23.106490 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 15:57:23 crc kubenswrapper[4894]: I1209 15:57:23.525326 4894 generic.go:334] "Generic (PLEG): container finished" podID="e8d5cd4c-3ff6-4a99-84c9-f051224f09cd" containerID="719f8b0d41d934dcea4a3b5b0a1b392b95070bf217bffe7b83665a5a24925915" exitCode=0 Dec 09 15:57:23 crc kubenswrapper[4894]: I1209 15:57:23.525420 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swzkf" event={"ID":"e8d5cd4c-3ff6-4a99-84c9-f051224f09cd","Type":"ContainerDied","Data":"719f8b0d41d934dcea4a3b5b0a1b392b95070bf217bffe7b83665a5a24925915"} Dec 09 15:57:23 crc kubenswrapper[4894]: I1209 15:57:23.525661 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swzkf" event={"ID":"e8d5cd4c-3ff6-4a99-84c9-f051224f09cd","Type":"ContainerStarted","Data":"30a0a22be584eaa3857e5523262c73024dd6b1baaac2ee4b1e4c452efb89bd1e"} Dec 09 15:57:23 crc kubenswrapper[4894]: I1209 15:57:23.528064 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 15:57:25 crc kubenswrapper[4894]: I1209 15:57:25.547393 4894 generic.go:334] "Generic (PLEG): container finished" podID="e8d5cd4c-3ff6-4a99-84c9-f051224f09cd" containerID="f310d0bbeffa3d5326fa7e2c92ba05f56f4d2a6346735d3235d2597fefab3f20" exitCode=0 Dec 09 15:57:25 crc kubenswrapper[4894]: I1209 15:57:25.547473 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swzkf" event={"ID":"e8d5cd4c-3ff6-4a99-84c9-f051224f09cd","Type":"ContainerDied","Data":"f310d0bbeffa3d5326fa7e2c92ba05f56f4d2a6346735d3235d2597fefab3f20"} Dec 09 15:57:26 crc kubenswrapper[4894]: I1209 15:57:26.560628 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swzkf" event={"ID":"e8d5cd4c-3ff6-4a99-84c9-f051224f09cd","Type":"ContainerStarted","Data":"72fba8bdd9c6611e9ffefbe7a0850808136ec6b9e95210864da7edf5c117a09d"} Dec 09 15:57:26 crc kubenswrapper[4894]: I1209 15:57:26.583650 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-swzkf" podStartSLOduration=2.049924899 podStartE2EDuration="4.583619283s" podCreationTimestamp="2025-12-09 15:57:22 +0000 UTC" firstStartedPulling="2025-12-09 15:57:23.527820332 +0000 UTC m=+1537.847031011" lastFinishedPulling="2025-12-09 15:57:26.061514726 +0000 UTC m=+1540.380725395" observedRunningTime="2025-12-09 15:57:26.581309397 +0000 UTC m=+1540.900520066" watchObservedRunningTime="2025-12-09 15:57:26.583619283 +0000 UTC m=+1540.902829952" Dec 09 15:57:32 crc kubenswrapper[4894]: I1209 15:57:32.504892 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-swzkf" Dec 09 15:57:32 crc kubenswrapper[4894]: I1209 15:57:32.505473 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-swzkf" Dec 09 15:57:32 crc kubenswrapper[4894]: I1209 15:57:32.575876 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-swzkf" Dec 09 15:57:32 crc kubenswrapper[4894]: I1209 15:57:32.659669 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-swzkf" Dec 09 15:57:32 crc kubenswrapper[4894]: I1209 15:57:32.813793 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-swzkf"] Dec 09 15:57:34 crc kubenswrapper[4894]: I1209 15:57:34.626855 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-swzkf" podUID="e8d5cd4c-3ff6-4a99-84c9-f051224f09cd" containerName="registry-server" containerID="cri-o://72fba8bdd9c6611e9ffefbe7a0850808136ec6b9e95210864da7edf5c117a09d" gracePeriod=2 Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.168926 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swzkf" Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.339175 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8d5cd4c-3ff6-4a99-84c9-f051224f09cd-catalog-content\") pod \"e8d5cd4c-3ff6-4a99-84c9-f051224f09cd\" (UID: \"e8d5cd4c-3ff6-4a99-84c9-f051224f09cd\") " Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.339347 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwttm\" (UniqueName: \"kubernetes.io/projected/e8d5cd4c-3ff6-4a99-84c9-f051224f09cd-kube-api-access-fwttm\") pod \"e8d5cd4c-3ff6-4a99-84c9-f051224f09cd\" (UID: \"e8d5cd4c-3ff6-4a99-84c9-f051224f09cd\") " Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.339406 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8d5cd4c-3ff6-4a99-84c9-f051224f09cd-utilities\") pod \"e8d5cd4c-3ff6-4a99-84c9-f051224f09cd\" (UID: \"e8d5cd4c-3ff6-4a99-84c9-f051224f09cd\") " Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.342850 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8d5cd4c-3ff6-4a99-84c9-f051224f09cd-utilities" (OuterVolumeSpecName: "utilities") pod "e8d5cd4c-3ff6-4a99-84c9-f051224f09cd" (UID: "e8d5cd4c-3ff6-4a99-84c9-f051224f09cd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.362091 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8d5cd4c-3ff6-4a99-84c9-f051224f09cd-kube-api-access-fwttm" (OuterVolumeSpecName: "kube-api-access-fwttm") pod "e8d5cd4c-3ff6-4a99-84c9-f051224f09cd" (UID: "e8d5cd4c-3ff6-4a99-84c9-f051224f09cd"). InnerVolumeSpecName "kube-api-access-fwttm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.410687 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8d5cd4c-3ff6-4a99-84c9-f051224f09cd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e8d5cd4c-3ff6-4a99-84c9-f051224f09cd" (UID: "e8d5cd4c-3ff6-4a99-84c9-f051224f09cd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.442479 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e8d5cd4c-3ff6-4a99-84c9-f051224f09cd-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.442510 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwttm\" (UniqueName: \"kubernetes.io/projected/e8d5cd4c-3ff6-4a99-84c9-f051224f09cd-kube-api-access-fwttm\") on node \"crc\" DevicePath \"\"" Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.442522 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e8d5cd4c-3ff6-4a99-84c9-f051224f09cd-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.649360 4894 generic.go:334] "Generic (PLEG): container finished" podID="e8d5cd4c-3ff6-4a99-84c9-f051224f09cd" containerID="72fba8bdd9c6611e9ffefbe7a0850808136ec6b9e95210864da7edf5c117a09d" exitCode=0 Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.649415 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swzkf" event={"ID":"e8d5cd4c-3ff6-4a99-84c9-f051224f09cd","Type":"ContainerDied","Data":"72fba8bdd9c6611e9ffefbe7a0850808136ec6b9e95210864da7edf5c117a09d"} Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.649448 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-swzkf" event={"ID":"e8d5cd4c-3ff6-4a99-84c9-f051224f09cd","Type":"ContainerDied","Data":"30a0a22be584eaa3857e5523262c73024dd6b1baaac2ee4b1e4c452efb89bd1e"} Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.649469 4894 scope.go:117] "RemoveContainer" containerID="72fba8bdd9c6611e9ffefbe7a0850808136ec6b9e95210864da7edf5c117a09d" Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.649658 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-swzkf" Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.690111 4894 scope.go:117] "RemoveContainer" containerID="f310d0bbeffa3d5326fa7e2c92ba05f56f4d2a6346735d3235d2597fefab3f20" Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.724704 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-swzkf"] Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.737285 4894 scope.go:117] "RemoveContainer" containerID="719f8b0d41d934dcea4a3b5b0a1b392b95070bf217bffe7b83665a5a24925915" Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.754938 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-swzkf"] Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.815893 4894 scope.go:117] "RemoveContainer" containerID="72fba8bdd9c6611e9ffefbe7a0850808136ec6b9e95210864da7edf5c117a09d" Dec 09 15:57:35 crc kubenswrapper[4894]: E1209 15:57:35.816404 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"72fba8bdd9c6611e9ffefbe7a0850808136ec6b9e95210864da7edf5c117a09d\": container with ID starting with 72fba8bdd9c6611e9ffefbe7a0850808136ec6b9e95210864da7edf5c117a09d not found: ID does not exist" containerID="72fba8bdd9c6611e9ffefbe7a0850808136ec6b9e95210864da7edf5c117a09d" Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.816462 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"72fba8bdd9c6611e9ffefbe7a0850808136ec6b9e95210864da7edf5c117a09d"} err="failed to get container status \"72fba8bdd9c6611e9ffefbe7a0850808136ec6b9e95210864da7edf5c117a09d\": rpc error: code = NotFound desc = could not find container \"72fba8bdd9c6611e9ffefbe7a0850808136ec6b9e95210864da7edf5c117a09d\": container with ID starting with 72fba8bdd9c6611e9ffefbe7a0850808136ec6b9e95210864da7edf5c117a09d not found: ID does not exist" Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.816486 4894 scope.go:117] "RemoveContainer" containerID="f310d0bbeffa3d5326fa7e2c92ba05f56f4d2a6346735d3235d2597fefab3f20" Dec 09 15:57:35 crc kubenswrapper[4894]: E1209 15:57:35.817323 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f310d0bbeffa3d5326fa7e2c92ba05f56f4d2a6346735d3235d2597fefab3f20\": container with ID starting with f310d0bbeffa3d5326fa7e2c92ba05f56f4d2a6346735d3235d2597fefab3f20 not found: ID does not exist" containerID="f310d0bbeffa3d5326fa7e2c92ba05f56f4d2a6346735d3235d2597fefab3f20" Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.817355 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f310d0bbeffa3d5326fa7e2c92ba05f56f4d2a6346735d3235d2597fefab3f20"} err="failed to get container status \"f310d0bbeffa3d5326fa7e2c92ba05f56f4d2a6346735d3235d2597fefab3f20\": rpc error: code = NotFound desc = could not find container \"f310d0bbeffa3d5326fa7e2c92ba05f56f4d2a6346735d3235d2597fefab3f20\": container with ID starting with f310d0bbeffa3d5326fa7e2c92ba05f56f4d2a6346735d3235d2597fefab3f20 not found: ID does not exist" Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.817377 4894 scope.go:117] "RemoveContainer" containerID="719f8b0d41d934dcea4a3b5b0a1b392b95070bf217bffe7b83665a5a24925915" Dec 09 15:57:35 crc kubenswrapper[4894]: E1209 15:57:35.817679 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"719f8b0d41d934dcea4a3b5b0a1b392b95070bf217bffe7b83665a5a24925915\": container with ID starting with 719f8b0d41d934dcea4a3b5b0a1b392b95070bf217bffe7b83665a5a24925915 not found: ID does not exist" containerID="719f8b0d41d934dcea4a3b5b0a1b392b95070bf217bffe7b83665a5a24925915" Dec 09 15:57:35 crc kubenswrapper[4894]: I1209 15:57:35.817764 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"719f8b0d41d934dcea4a3b5b0a1b392b95070bf217bffe7b83665a5a24925915"} err="failed to get container status \"719f8b0d41d934dcea4a3b5b0a1b392b95070bf217bffe7b83665a5a24925915\": rpc error: code = NotFound desc = could not find container \"719f8b0d41d934dcea4a3b5b0a1b392b95070bf217bffe7b83665a5a24925915\": container with ID starting with 719f8b0d41d934dcea4a3b5b0a1b392b95070bf217bffe7b83665a5a24925915 not found: ID does not exist" Dec 09 15:57:36 crc kubenswrapper[4894]: I1209 15:57:36.116979 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 15:57:36 crc kubenswrapper[4894]: E1209 15:57:36.117229 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 15:57:36 crc kubenswrapper[4894]: I1209 15:57:36.123262 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8d5cd4c-3ff6-4a99-84c9-f051224f09cd" path="/var/lib/kubelet/pods/e8d5cd4c-3ff6-4a99-84c9-f051224f09cd/volumes" Dec 09 15:57:51 crc kubenswrapper[4894]: I1209 15:57:51.106254 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 15:57:51 crc kubenswrapper[4894]: E1209 15:57:51.107342 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 15:57:56 crc kubenswrapper[4894]: I1209 15:57:56.880185 4894 generic.go:334] "Generic (PLEG): container finished" podID="be481ff5-0866-4a78-84a5-2963c405dd12" containerID="6df01dbe7ee500afeee4fdf4e66c4836d9a37ef755c3d2a40fc7b6646637f264" exitCode=0 Dec 09 15:57:56 crc kubenswrapper[4894]: I1209 15:57:56.880320 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh" event={"ID":"be481ff5-0866-4a78-84a5-2963c405dd12","Type":"ContainerDied","Data":"6df01dbe7ee500afeee4fdf4e66c4836d9a37ef755c3d2a40fc7b6646637f264"} Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.383081 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh" Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.514128 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be481ff5-0866-4a78-84a5-2963c405dd12-ssh-key\") pod \"be481ff5-0866-4a78-84a5-2963c405dd12\" (UID: \"be481ff5-0866-4a78-84a5-2963c405dd12\") " Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.514242 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kqrn8\" (UniqueName: \"kubernetes.io/projected/be481ff5-0866-4a78-84a5-2963c405dd12-kube-api-access-kqrn8\") pod \"be481ff5-0866-4a78-84a5-2963c405dd12\" (UID: \"be481ff5-0866-4a78-84a5-2963c405dd12\") " Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.514345 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be481ff5-0866-4a78-84a5-2963c405dd12-inventory\") pod \"be481ff5-0866-4a78-84a5-2963c405dd12\" (UID: \"be481ff5-0866-4a78-84a5-2963c405dd12\") " Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.514486 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be481ff5-0866-4a78-84a5-2963c405dd12-bootstrap-combined-ca-bundle\") pod \"be481ff5-0866-4a78-84a5-2963c405dd12\" (UID: \"be481ff5-0866-4a78-84a5-2963c405dd12\") " Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.520594 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be481ff5-0866-4a78-84a5-2963c405dd12-kube-api-access-kqrn8" (OuterVolumeSpecName: "kube-api-access-kqrn8") pod "be481ff5-0866-4a78-84a5-2963c405dd12" (UID: "be481ff5-0866-4a78-84a5-2963c405dd12"). InnerVolumeSpecName "kube-api-access-kqrn8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.522477 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be481ff5-0866-4a78-84a5-2963c405dd12-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "be481ff5-0866-4a78-84a5-2963c405dd12" (UID: "be481ff5-0866-4a78-84a5-2963c405dd12"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.561151 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be481ff5-0866-4a78-84a5-2963c405dd12-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "be481ff5-0866-4a78-84a5-2963c405dd12" (UID: "be481ff5-0866-4a78-84a5-2963c405dd12"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.566271 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be481ff5-0866-4a78-84a5-2963c405dd12-inventory" (OuterVolumeSpecName: "inventory") pod "be481ff5-0866-4a78-84a5-2963c405dd12" (UID: "be481ff5-0866-4a78-84a5-2963c405dd12"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.617152 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be481ff5-0866-4a78-84a5-2963c405dd12-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.617202 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kqrn8\" (UniqueName: \"kubernetes.io/projected/be481ff5-0866-4a78-84a5-2963c405dd12-kube-api-access-kqrn8\") on node \"crc\" DevicePath \"\"" Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.617224 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be481ff5-0866-4a78-84a5-2963c405dd12-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.617245 4894 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be481ff5-0866-4a78-84a5-2963c405dd12-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.931448 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh" event={"ID":"be481ff5-0866-4a78-84a5-2963c405dd12","Type":"ContainerDied","Data":"5bfa6afb217c5595218b1510eba3d255bc18d7e59ae4e1e77ee602a9f11245a2"} Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.931510 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5bfa6afb217c5595218b1510eba3d255bc18d7e59ae4e1e77ee602a9f11245a2" Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.931518 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh" Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.996652 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p"] Dec 09 15:57:58 crc kubenswrapper[4894]: E1209 15:57:58.997097 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8d5cd4c-3ff6-4a99-84c9-f051224f09cd" containerName="registry-server" Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.997122 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8d5cd4c-3ff6-4a99-84c9-f051224f09cd" containerName="registry-server" Dec 09 15:57:58 crc kubenswrapper[4894]: E1209 15:57:58.997135 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8d5cd4c-3ff6-4a99-84c9-f051224f09cd" containerName="extract-utilities" Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.997143 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8d5cd4c-3ff6-4a99-84c9-f051224f09cd" containerName="extract-utilities" Dec 09 15:57:58 crc kubenswrapper[4894]: E1209 15:57:58.997165 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8d5cd4c-3ff6-4a99-84c9-f051224f09cd" containerName="extract-content" Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.997174 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8d5cd4c-3ff6-4a99-84c9-f051224f09cd" containerName="extract-content" Dec 09 15:57:58 crc kubenswrapper[4894]: E1209 15:57:58.997186 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be481ff5-0866-4a78-84a5-2963c405dd12" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.997195 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="be481ff5-0866-4a78-84a5-2963c405dd12" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.997380 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8d5cd4c-3ff6-4a99-84c9-f051224f09cd" containerName="registry-server" Dec 09 15:57:58 crc kubenswrapper[4894]: I1209 15:57:58.997394 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="be481ff5-0866-4a78-84a5-2963c405dd12" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 09 15:57:59 crc kubenswrapper[4894]: I1209 15:57:59.002229 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p" Dec 09 15:57:59 crc kubenswrapper[4894]: I1209 15:57:59.005792 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 15:57:59 crc kubenswrapper[4894]: I1209 15:57:59.005968 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 15:57:59 crc kubenswrapper[4894]: I1209 15:57:59.006194 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 15:57:59 crc kubenswrapper[4894]: I1209 15:57:59.006402 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 15:57:59 crc kubenswrapper[4894]: I1209 15:57:59.013627 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p"] Dec 09 15:57:59 crc kubenswrapper[4894]: I1209 15:57:59.136053 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4p48g\" (UniqueName: \"kubernetes.io/projected/db9fb459-0a0a-4a5f-8510-12adae609c7f-kube-api-access-4p48g\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-97l8p\" (UID: \"db9fb459-0a0a-4a5f-8510-12adae609c7f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p" Dec 09 15:57:59 crc kubenswrapper[4894]: I1209 15:57:59.136129 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db9fb459-0a0a-4a5f-8510-12adae609c7f-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-97l8p\" (UID: \"db9fb459-0a0a-4a5f-8510-12adae609c7f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p" Dec 09 15:57:59 crc kubenswrapper[4894]: I1209 15:57:59.136165 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db9fb459-0a0a-4a5f-8510-12adae609c7f-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-97l8p\" (UID: \"db9fb459-0a0a-4a5f-8510-12adae609c7f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p" Dec 09 15:57:59 crc kubenswrapper[4894]: I1209 15:57:59.239425 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db9fb459-0a0a-4a5f-8510-12adae609c7f-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-97l8p\" (UID: \"db9fb459-0a0a-4a5f-8510-12adae609c7f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p" Dec 09 15:57:59 crc kubenswrapper[4894]: I1209 15:57:59.239934 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db9fb459-0a0a-4a5f-8510-12adae609c7f-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-97l8p\" (UID: \"db9fb459-0a0a-4a5f-8510-12adae609c7f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p" Dec 09 15:57:59 crc kubenswrapper[4894]: I1209 15:57:59.240169 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4p48g\" (UniqueName: \"kubernetes.io/projected/db9fb459-0a0a-4a5f-8510-12adae609c7f-kube-api-access-4p48g\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-97l8p\" (UID: \"db9fb459-0a0a-4a5f-8510-12adae609c7f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p" Dec 09 15:57:59 crc kubenswrapper[4894]: I1209 15:57:59.244533 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db9fb459-0a0a-4a5f-8510-12adae609c7f-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-97l8p\" (UID: \"db9fb459-0a0a-4a5f-8510-12adae609c7f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p" Dec 09 15:57:59 crc kubenswrapper[4894]: I1209 15:57:59.247859 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db9fb459-0a0a-4a5f-8510-12adae609c7f-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-97l8p\" (UID: \"db9fb459-0a0a-4a5f-8510-12adae609c7f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p" Dec 09 15:57:59 crc kubenswrapper[4894]: I1209 15:57:59.283091 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4p48g\" (UniqueName: \"kubernetes.io/projected/db9fb459-0a0a-4a5f-8510-12adae609c7f-kube-api-access-4p48g\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-97l8p\" (UID: \"db9fb459-0a0a-4a5f-8510-12adae609c7f\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p" Dec 09 15:57:59 crc kubenswrapper[4894]: I1209 15:57:59.326769 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p" Dec 09 15:57:59 crc kubenswrapper[4894]: I1209 15:57:59.891283 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p"] Dec 09 15:57:59 crc kubenswrapper[4894]: I1209 15:57:59.944042 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p" event={"ID":"db9fb459-0a0a-4a5f-8510-12adae609c7f","Type":"ContainerStarted","Data":"fab29074615c1463257fdb43179229f2cc8f508ffe56b39ef48e35c107e9ae90"} Dec 09 15:58:01 crc kubenswrapper[4894]: I1209 15:58:01.962159 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p" event={"ID":"db9fb459-0a0a-4a5f-8510-12adae609c7f","Type":"ContainerStarted","Data":"55c85647833560d5e7e7e230a2b12b9f0ed5bce0cb39436f7e3ea135e2061620"} Dec 09 15:58:01 crc kubenswrapper[4894]: I1209 15:58:01.982770 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p" podStartSLOduration=2.92491653 podStartE2EDuration="3.982741257s" podCreationTimestamp="2025-12-09 15:57:58 +0000 UTC" firstStartedPulling="2025-12-09 15:57:59.890856247 +0000 UTC m=+1574.210066916" lastFinishedPulling="2025-12-09 15:58:00.948680964 +0000 UTC m=+1575.267891643" observedRunningTime="2025-12-09 15:58:01.978816981 +0000 UTC m=+1576.298027640" watchObservedRunningTime="2025-12-09 15:58:01.982741257 +0000 UTC m=+1576.301951936" Dec 09 15:58:06 crc kubenswrapper[4894]: I1209 15:58:06.107504 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 15:58:06 crc kubenswrapper[4894]: E1209 15:58:06.112701 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 15:58:07 crc kubenswrapper[4894]: I1209 15:58:07.053171 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-hp2h5"] Dec 09 15:58:07 crc kubenswrapper[4894]: I1209 15:58:07.060680 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-hp2h5"] Dec 09 15:58:07 crc kubenswrapper[4894]: I1209 15:58:07.069979 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-11cf-account-create-update-ggv4k"] Dec 09 15:58:07 crc kubenswrapper[4894]: I1209 15:58:07.078108 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-11cf-account-create-update-ggv4k"] Dec 09 15:58:08 crc kubenswrapper[4894]: I1209 15:58:08.116565 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d962c43-b57b-4db6-bda2-c6b48bb7335a" path="/var/lib/kubelet/pods/9d962c43-b57b-4db6-bda2-c6b48bb7335a/volumes" Dec 09 15:58:08 crc kubenswrapper[4894]: I1209 15:58:08.117422 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b" path="/var/lib/kubelet/pods/f6e6dd91-2ce9-4c86-8ea7-a06c42d84b0b/volumes" Dec 09 15:58:10 crc kubenswrapper[4894]: I1209 15:58:10.706092 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-b5czq"] Dec 09 15:58:10 crc kubenswrapper[4894]: I1209 15:58:10.708485 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b5czq" Dec 09 15:58:10 crc kubenswrapper[4894]: I1209 15:58:10.716125 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b5czq"] Dec 09 15:58:10 crc kubenswrapper[4894]: I1209 15:58:10.856980 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48095069-073b-4ea6-9608-2723fae94f24-utilities\") pod \"redhat-marketplace-b5czq\" (UID: \"48095069-073b-4ea6-9608-2723fae94f24\") " pod="openshift-marketplace/redhat-marketplace-b5czq" Dec 09 15:58:10 crc kubenswrapper[4894]: I1209 15:58:10.857052 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48095069-073b-4ea6-9608-2723fae94f24-catalog-content\") pod \"redhat-marketplace-b5czq\" (UID: \"48095069-073b-4ea6-9608-2723fae94f24\") " pod="openshift-marketplace/redhat-marketplace-b5czq" Dec 09 15:58:10 crc kubenswrapper[4894]: I1209 15:58:10.857570 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cws6w\" (UniqueName: \"kubernetes.io/projected/48095069-073b-4ea6-9608-2723fae94f24-kube-api-access-cws6w\") pod \"redhat-marketplace-b5czq\" (UID: \"48095069-073b-4ea6-9608-2723fae94f24\") " pod="openshift-marketplace/redhat-marketplace-b5czq" Dec 09 15:58:10 crc kubenswrapper[4894]: I1209 15:58:10.959044 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cws6w\" (UniqueName: \"kubernetes.io/projected/48095069-073b-4ea6-9608-2723fae94f24-kube-api-access-cws6w\") pod \"redhat-marketplace-b5czq\" (UID: \"48095069-073b-4ea6-9608-2723fae94f24\") " pod="openshift-marketplace/redhat-marketplace-b5czq" Dec 09 15:58:10 crc kubenswrapper[4894]: I1209 15:58:10.959138 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48095069-073b-4ea6-9608-2723fae94f24-utilities\") pod \"redhat-marketplace-b5czq\" (UID: \"48095069-073b-4ea6-9608-2723fae94f24\") " pod="openshift-marketplace/redhat-marketplace-b5czq" Dec 09 15:58:10 crc kubenswrapper[4894]: I1209 15:58:10.959194 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48095069-073b-4ea6-9608-2723fae94f24-catalog-content\") pod \"redhat-marketplace-b5czq\" (UID: \"48095069-073b-4ea6-9608-2723fae94f24\") " pod="openshift-marketplace/redhat-marketplace-b5czq" Dec 09 15:58:10 crc kubenswrapper[4894]: I1209 15:58:10.959649 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48095069-073b-4ea6-9608-2723fae94f24-utilities\") pod \"redhat-marketplace-b5czq\" (UID: \"48095069-073b-4ea6-9608-2723fae94f24\") " pod="openshift-marketplace/redhat-marketplace-b5czq" Dec 09 15:58:10 crc kubenswrapper[4894]: I1209 15:58:10.959684 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48095069-073b-4ea6-9608-2723fae94f24-catalog-content\") pod \"redhat-marketplace-b5czq\" (UID: \"48095069-073b-4ea6-9608-2723fae94f24\") " pod="openshift-marketplace/redhat-marketplace-b5czq" Dec 09 15:58:10 crc kubenswrapper[4894]: I1209 15:58:10.988111 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cws6w\" (UniqueName: \"kubernetes.io/projected/48095069-073b-4ea6-9608-2723fae94f24-kube-api-access-cws6w\") pod \"redhat-marketplace-b5czq\" (UID: \"48095069-073b-4ea6-9608-2723fae94f24\") " pod="openshift-marketplace/redhat-marketplace-b5czq" Dec 09 15:58:11 crc kubenswrapper[4894]: I1209 15:58:11.034740 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-9j5zq"] Dec 09 15:58:11 crc kubenswrapper[4894]: I1209 15:58:11.043554 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-9j5zq"] Dec 09 15:58:11 crc kubenswrapper[4894]: I1209 15:58:11.057209 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b5czq" Dec 09 15:58:11 crc kubenswrapper[4894]: I1209 15:58:11.526507 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-b5czq"] Dec 09 15:58:11 crc kubenswrapper[4894]: W1209 15:58:11.542779 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod48095069_073b_4ea6_9608_2723fae94f24.slice/crio-8780f0c8d60eb64d3ba0dce2a11fa7d1fbe5c0b2c05f5105d0afee5a6ca17fbf WatchSource:0}: Error finding container 8780f0c8d60eb64d3ba0dce2a11fa7d1fbe5c0b2c05f5105d0afee5a6ca17fbf: Status 404 returned error can't find the container with id 8780f0c8d60eb64d3ba0dce2a11fa7d1fbe5c0b2c05f5105d0afee5a6ca17fbf Dec 09 15:58:12 crc kubenswrapper[4894]: I1209 15:58:12.035211 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-8cc8-account-create-update-8f2ql"] Dec 09 15:58:12 crc kubenswrapper[4894]: I1209 15:58:12.050742 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-8cc8-account-create-update-8f2ql"] Dec 09 15:58:12 crc kubenswrapper[4894]: I1209 15:58:12.096017 4894 generic.go:334] "Generic (PLEG): container finished" podID="48095069-073b-4ea6-9608-2723fae94f24" containerID="6ac33a37d1d3a00a2ca0d1ae598fe1170e3272016159bfe73cc1c9ce33099f11" exitCode=0 Dec 09 15:58:12 crc kubenswrapper[4894]: I1209 15:58:12.096063 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5czq" event={"ID":"48095069-073b-4ea6-9608-2723fae94f24","Type":"ContainerDied","Data":"6ac33a37d1d3a00a2ca0d1ae598fe1170e3272016159bfe73cc1c9ce33099f11"} Dec 09 15:58:12 crc kubenswrapper[4894]: I1209 15:58:12.096098 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5czq" event={"ID":"48095069-073b-4ea6-9608-2723fae94f24","Type":"ContainerStarted","Data":"8780f0c8d60eb64d3ba0dce2a11fa7d1fbe5c0b2c05f5105d0afee5a6ca17fbf"} Dec 09 15:58:12 crc kubenswrapper[4894]: I1209 15:58:12.127749 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05b5bde5-3e09-4764-b41f-0de53ad5207b" path="/var/lib/kubelet/pods/05b5bde5-3e09-4764-b41f-0de53ad5207b/volumes" Dec 09 15:58:12 crc kubenswrapper[4894]: I1209 15:58:12.129176 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9b968921-706b-4a6d-a667-2cc3389710ef" path="/var/lib/kubelet/pods/9b968921-706b-4a6d-a667-2cc3389710ef/volumes" Dec 09 15:58:13 crc kubenswrapper[4894]: I1209 15:58:13.039722 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-1747-account-create-update-6fqrf"] Dec 09 15:58:13 crc kubenswrapper[4894]: I1209 15:58:13.053096 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-tqxhm"] Dec 09 15:58:13 crc kubenswrapper[4894]: I1209 15:58:13.069896 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-1747-account-create-update-6fqrf"] Dec 09 15:58:13 crc kubenswrapper[4894]: I1209 15:58:13.078939 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-tqxhm"] Dec 09 15:58:13 crc kubenswrapper[4894]: I1209 15:58:13.109901 4894 generic.go:334] "Generic (PLEG): container finished" podID="48095069-073b-4ea6-9608-2723fae94f24" containerID="cb88f65109bb689af754afe2454d683decc78dd9877ac0647818d03382767247" exitCode=0 Dec 09 15:58:13 crc kubenswrapper[4894]: I1209 15:58:13.109983 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5czq" event={"ID":"48095069-073b-4ea6-9608-2723fae94f24","Type":"ContainerDied","Data":"cb88f65109bb689af754afe2454d683decc78dd9877ac0647818d03382767247"} Dec 09 15:58:14 crc kubenswrapper[4894]: I1209 15:58:14.121893 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1" path="/var/lib/kubelet/pods/372381bb-d5ff-47d7-a2bb-9ba8cfeacbe1/volumes" Dec 09 15:58:14 crc kubenswrapper[4894]: I1209 15:58:14.123207 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dc5ff393-b30e-448b-a618-09a3b3162c97" path="/var/lib/kubelet/pods/dc5ff393-b30e-448b-a618-09a3b3162c97/volumes" Dec 09 15:58:14 crc kubenswrapper[4894]: I1209 15:58:14.123832 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5czq" event={"ID":"48095069-073b-4ea6-9608-2723fae94f24","Type":"ContainerStarted","Data":"7d93d30d869d18a07b78642e5db5ad109e11c3b1bfeb34f27cdf10b044217b6b"} Dec 09 15:58:14 crc kubenswrapper[4894]: I1209 15:58:14.147600 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-b5czq" podStartSLOduration=2.620119809 podStartE2EDuration="4.147580658s" podCreationTimestamp="2025-12-09 15:58:10 +0000 UTC" firstStartedPulling="2025-12-09 15:58:12.099734847 +0000 UTC m=+1586.418945556" lastFinishedPulling="2025-12-09 15:58:13.627195736 +0000 UTC m=+1587.946406405" observedRunningTime="2025-12-09 15:58:14.139379836 +0000 UTC m=+1588.458590545" watchObservedRunningTime="2025-12-09 15:58:14.147580658 +0000 UTC m=+1588.466791337" Dec 09 15:58:20 crc kubenswrapper[4894]: I1209 15:58:20.107707 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 15:58:20 crc kubenswrapper[4894]: E1209 15:58:20.108481 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 15:58:21 crc kubenswrapper[4894]: I1209 15:58:21.057680 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-b5czq" Dec 09 15:58:21 crc kubenswrapper[4894]: I1209 15:58:21.057737 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-b5czq" Dec 09 15:58:21 crc kubenswrapper[4894]: I1209 15:58:21.105531 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-b5czq" Dec 09 15:58:21 crc kubenswrapper[4894]: I1209 15:58:21.228759 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-b5czq" Dec 09 15:58:21 crc kubenswrapper[4894]: I1209 15:58:21.345067 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b5czq"] Dec 09 15:58:23 crc kubenswrapper[4894]: I1209 15:58:23.201740 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-b5czq" podUID="48095069-073b-4ea6-9608-2723fae94f24" containerName="registry-server" containerID="cri-o://7d93d30d869d18a07b78642e5db5ad109e11c3b1bfeb34f27cdf10b044217b6b" gracePeriod=2 Dec 09 15:58:23 crc kubenswrapper[4894]: I1209 15:58:23.825389 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b5czq" Dec 09 15:58:23 crc kubenswrapper[4894]: I1209 15:58:23.938799 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cws6w\" (UniqueName: \"kubernetes.io/projected/48095069-073b-4ea6-9608-2723fae94f24-kube-api-access-cws6w\") pod \"48095069-073b-4ea6-9608-2723fae94f24\" (UID: \"48095069-073b-4ea6-9608-2723fae94f24\") " Dec 09 15:58:23 crc kubenswrapper[4894]: I1209 15:58:23.939098 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48095069-073b-4ea6-9608-2723fae94f24-catalog-content\") pod \"48095069-073b-4ea6-9608-2723fae94f24\" (UID: \"48095069-073b-4ea6-9608-2723fae94f24\") " Dec 09 15:58:23 crc kubenswrapper[4894]: I1209 15:58:23.939190 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48095069-073b-4ea6-9608-2723fae94f24-utilities\") pod \"48095069-073b-4ea6-9608-2723fae94f24\" (UID: \"48095069-073b-4ea6-9608-2723fae94f24\") " Dec 09 15:58:23 crc kubenswrapper[4894]: I1209 15:58:23.940328 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48095069-073b-4ea6-9608-2723fae94f24-utilities" (OuterVolumeSpecName: "utilities") pod "48095069-073b-4ea6-9608-2723fae94f24" (UID: "48095069-073b-4ea6-9608-2723fae94f24"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:58:23 crc kubenswrapper[4894]: I1209 15:58:23.944553 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/48095069-073b-4ea6-9608-2723fae94f24-kube-api-access-cws6w" (OuterVolumeSpecName: "kube-api-access-cws6w") pod "48095069-073b-4ea6-9608-2723fae94f24" (UID: "48095069-073b-4ea6-9608-2723fae94f24"). InnerVolumeSpecName "kube-api-access-cws6w". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:58:23 crc kubenswrapper[4894]: I1209 15:58:23.959104 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/48095069-073b-4ea6-9608-2723fae94f24-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "48095069-073b-4ea6-9608-2723fae94f24" (UID: "48095069-073b-4ea6-9608-2723fae94f24"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 15:58:24 crc kubenswrapper[4894]: I1209 15:58:24.041541 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/48095069-073b-4ea6-9608-2723fae94f24-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 15:58:24 crc kubenswrapper[4894]: I1209 15:58:24.041907 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/48095069-073b-4ea6-9608-2723fae94f24-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 15:58:24 crc kubenswrapper[4894]: I1209 15:58:24.041921 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cws6w\" (UniqueName: \"kubernetes.io/projected/48095069-073b-4ea6-9608-2723fae94f24-kube-api-access-cws6w\") on node \"crc\" DevicePath \"\"" Dec 09 15:58:24 crc kubenswrapper[4894]: I1209 15:58:24.216597 4894 generic.go:334] "Generic (PLEG): container finished" podID="48095069-073b-4ea6-9608-2723fae94f24" containerID="7d93d30d869d18a07b78642e5db5ad109e11c3b1bfeb34f27cdf10b044217b6b" exitCode=0 Dec 09 15:58:24 crc kubenswrapper[4894]: I1209 15:58:24.216714 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-b5czq" Dec 09 15:58:24 crc kubenswrapper[4894]: I1209 15:58:24.216716 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5czq" event={"ID":"48095069-073b-4ea6-9608-2723fae94f24","Type":"ContainerDied","Data":"7d93d30d869d18a07b78642e5db5ad109e11c3b1bfeb34f27cdf10b044217b6b"} Dec 09 15:58:24 crc kubenswrapper[4894]: I1209 15:58:24.216813 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-b5czq" event={"ID":"48095069-073b-4ea6-9608-2723fae94f24","Type":"ContainerDied","Data":"8780f0c8d60eb64d3ba0dce2a11fa7d1fbe5c0b2c05f5105d0afee5a6ca17fbf"} Dec 09 15:58:24 crc kubenswrapper[4894]: I1209 15:58:24.216846 4894 scope.go:117] "RemoveContainer" containerID="7d93d30d869d18a07b78642e5db5ad109e11c3b1bfeb34f27cdf10b044217b6b" Dec 09 15:58:24 crc kubenswrapper[4894]: I1209 15:58:24.256307 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-b5czq"] Dec 09 15:58:24 crc kubenswrapper[4894]: I1209 15:58:24.268600 4894 scope.go:117] "RemoveContainer" containerID="cb88f65109bb689af754afe2454d683decc78dd9877ac0647818d03382767247" Dec 09 15:58:24 crc kubenswrapper[4894]: I1209 15:58:24.271179 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-b5czq"] Dec 09 15:58:24 crc kubenswrapper[4894]: I1209 15:58:24.294520 4894 scope.go:117] "RemoveContainer" containerID="6ac33a37d1d3a00a2ca0d1ae598fe1170e3272016159bfe73cc1c9ce33099f11" Dec 09 15:58:24 crc kubenswrapper[4894]: I1209 15:58:24.362740 4894 scope.go:117] "RemoveContainer" containerID="7d93d30d869d18a07b78642e5db5ad109e11c3b1bfeb34f27cdf10b044217b6b" Dec 09 15:58:24 crc kubenswrapper[4894]: E1209 15:58:24.363299 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7d93d30d869d18a07b78642e5db5ad109e11c3b1bfeb34f27cdf10b044217b6b\": container with ID starting with 7d93d30d869d18a07b78642e5db5ad109e11c3b1bfeb34f27cdf10b044217b6b not found: ID does not exist" containerID="7d93d30d869d18a07b78642e5db5ad109e11c3b1bfeb34f27cdf10b044217b6b" Dec 09 15:58:24 crc kubenswrapper[4894]: I1209 15:58:24.363351 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7d93d30d869d18a07b78642e5db5ad109e11c3b1bfeb34f27cdf10b044217b6b"} err="failed to get container status \"7d93d30d869d18a07b78642e5db5ad109e11c3b1bfeb34f27cdf10b044217b6b\": rpc error: code = NotFound desc = could not find container \"7d93d30d869d18a07b78642e5db5ad109e11c3b1bfeb34f27cdf10b044217b6b\": container with ID starting with 7d93d30d869d18a07b78642e5db5ad109e11c3b1bfeb34f27cdf10b044217b6b not found: ID does not exist" Dec 09 15:58:24 crc kubenswrapper[4894]: I1209 15:58:24.363384 4894 scope.go:117] "RemoveContainer" containerID="cb88f65109bb689af754afe2454d683decc78dd9877ac0647818d03382767247" Dec 09 15:58:24 crc kubenswrapper[4894]: E1209 15:58:24.364464 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb88f65109bb689af754afe2454d683decc78dd9877ac0647818d03382767247\": container with ID starting with cb88f65109bb689af754afe2454d683decc78dd9877ac0647818d03382767247 not found: ID does not exist" containerID="cb88f65109bb689af754afe2454d683decc78dd9877ac0647818d03382767247" Dec 09 15:58:24 crc kubenswrapper[4894]: I1209 15:58:24.364517 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb88f65109bb689af754afe2454d683decc78dd9877ac0647818d03382767247"} err="failed to get container status \"cb88f65109bb689af754afe2454d683decc78dd9877ac0647818d03382767247\": rpc error: code = NotFound desc = could not find container \"cb88f65109bb689af754afe2454d683decc78dd9877ac0647818d03382767247\": container with ID starting with cb88f65109bb689af754afe2454d683decc78dd9877ac0647818d03382767247 not found: ID does not exist" Dec 09 15:58:24 crc kubenswrapper[4894]: I1209 15:58:24.364552 4894 scope.go:117] "RemoveContainer" containerID="6ac33a37d1d3a00a2ca0d1ae598fe1170e3272016159bfe73cc1c9ce33099f11" Dec 09 15:58:24 crc kubenswrapper[4894]: E1209 15:58:24.364870 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6ac33a37d1d3a00a2ca0d1ae598fe1170e3272016159bfe73cc1c9ce33099f11\": container with ID starting with 6ac33a37d1d3a00a2ca0d1ae598fe1170e3272016159bfe73cc1c9ce33099f11 not found: ID does not exist" containerID="6ac33a37d1d3a00a2ca0d1ae598fe1170e3272016159bfe73cc1c9ce33099f11" Dec 09 15:58:24 crc kubenswrapper[4894]: I1209 15:58:24.364902 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6ac33a37d1d3a00a2ca0d1ae598fe1170e3272016159bfe73cc1c9ce33099f11"} err="failed to get container status \"6ac33a37d1d3a00a2ca0d1ae598fe1170e3272016159bfe73cc1c9ce33099f11\": rpc error: code = NotFound desc = could not find container \"6ac33a37d1d3a00a2ca0d1ae598fe1170e3272016159bfe73cc1c9ce33099f11\": container with ID starting with 6ac33a37d1d3a00a2ca0d1ae598fe1170e3272016159bfe73cc1c9ce33099f11 not found: ID does not exist" Dec 09 15:58:26 crc kubenswrapper[4894]: I1209 15:58:26.119108 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48095069-073b-4ea6-9608-2723fae94f24" path="/var/lib/kubelet/pods/48095069-073b-4ea6-9608-2723fae94f24/volumes" Dec 09 15:58:31 crc kubenswrapper[4894]: I1209 15:58:31.045815 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-h74cz"] Dec 09 15:58:31 crc kubenswrapper[4894]: I1209 15:58:31.056004 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-h74cz"] Dec 09 15:58:32 crc kubenswrapper[4894]: I1209 15:58:32.106471 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 15:58:32 crc kubenswrapper[4894]: E1209 15:58:32.106814 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 15:58:32 crc kubenswrapper[4894]: I1209 15:58:32.120400 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26cc6eda-fee8-4183-b613-327cc2fad4b6" path="/var/lib/kubelet/pods/26cc6eda-fee8-4183-b613-327cc2fad4b6/volumes" Dec 09 15:58:39 crc kubenswrapper[4894]: I1209 15:58:39.044130 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-a57d-account-create-update-8xxg4"] Dec 09 15:58:39 crc kubenswrapper[4894]: I1209 15:58:39.062332 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-a57d-account-create-update-8xxg4"] Dec 09 15:58:40 crc kubenswrapper[4894]: I1209 15:58:40.049337 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-wpqxn"] Dec 09 15:58:40 crc kubenswrapper[4894]: I1209 15:58:40.065006 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-5da2-account-create-update-wthfq"] Dec 09 15:58:40 crc kubenswrapper[4894]: I1209 15:58:40.074166 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-v6zx2"] Dec 09 15:58:40 crc kubenswrapper[4894]: I1209 15:58:40.080572 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-wpqxn"] Dec 09 15:58:40 crc kubenswrapper[4894]: I1209 15:58:40.088963 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-fc79-account-create-update-ts2zw"] Dec 09 15:58:40 crc kubenswrapper[4894]: I1209 15:58:40.099003 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-pfn9g"] Dec 09 15:58:40 crc kubenswrapper[4894]: I1209 15:58:40.133112 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09b8ca6e-7521-4e7f-8272-504d2d7d66d7" path="/var/lib/kubelet/pods/09b8ca6e-7521-4e7f-8272-504d2d7d66d7/volumes" Dec 09 15:58:40 crc kubenswrapper[4894]: I1209 15:58:40.135519 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b191efc2-e3bf-49e8-83d9-7b2e55264863" path="/var/lib/kubelet/pods/b191efc2-e3bf-49e8-83d9-7b2e55264863/volumes" Dec 09 15:58:40 crc kubenswrapper[4894]: I1209 15:58:40.136059 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-fc79-account-create-update-ts2zw"] Dec 09 15:58:40 crc kubenswrapper[4894]: I1209 15:58:40.136086 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-pfn9g"] Dec 09 15:58:40 crc kubenswrapper[4894]: I1209 15:58:40.138658 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-v6zx2"] Dec 09 15:58:40 crc kubenswrapper[4894]: I1209 15:58:40.148490 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-5da2-account-create-update-wthfq"] Dec 09 15:58:42 crc kubenswrapper[4894]: I1209 15:58:42.121890 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="48cc59c5-cf8b-455b-8f27-beb7398510b3" path="/var/lib/kubelet/pods/48cc59c5-cf8b-455b-8f27-beb7398510b3/volumes" Dec 09 15:58:42 crc kubenswrapper[4894]: I1209 15:58:42.123736 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="81e9c778-7d46-47f9-8d63-15662adc4aa3" path="/var/lib/kubelet/pods/81e9c778-7d46-47f9-8d63-15662adc4aa3/volumes" Dec 09 15:58:42 crc kubenswrapper[4894]: I1209 15:58:42.124937 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="946604b5-e779-4579-a593-f46a7c4e3e5e" path="/var/lib/kubelet/pods/946604b5-e779-4579-a593-f46a7c4e3e5e/volumes" Dec 09 15:58:42 crc kubenswrapper[4894]: I1209 15:58:42.126158 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a754121b-d60b-4107-ac9c-15cc6db44d93" path="/var/lib/kubelet/pods/a754121b-d60b-4107-ac9c-15cc6db44d93/volumes" Dec 09 15:58:44 crc kubenswrapper[4894]: I1209 15:58:44.106202 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 15:58:44 crc kubenswrapper[4894]: E1209 15:58:44.106796 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 15:58:54 crc kubenswrapper[4894]: I1209 15:58:54.718527 4894 scope.go:117] "RemoveContainer" containerID="c690018c1d490040c1720247120a6a74843ecabd9c34e41fb997b014f699bf4c" Dec 09 15:58:54 crc kubenswrapper[4894]: I1209 15:58:54.752386 4894 scope.go:117] "RemoveContainer" containerID="723467964f41a9aa68ad1cbed609234ca51ea9e01d04854c81afa5b441925f83" Dec 09 15:58:54 crc kubenswrapper[4894]: I1209 15:58:54.801247 4894 scope.go:117] "RemoveContainer" containerID="2de20c72f91164109d228f1372091a3e7f4059758bc0f656b112da099e0f37d7" Dec 09 15:58:54 crc kubenswrapper[4894]: I1209 15:58:54.838581 4894 scope.go:117] "RemoveContainer" containerID="9d4e6632f32fe729acf3df190912cbb2f04261a364e68d5f5cde2931b3df2593" Dec 09 15:58:54 crc kubenswrapper[4894]: I1209 15:58:54.879393 4894 scope.go:117] "RemoveContainer" containerID="d66885ed9842006da476df221f882b744f4f5891d2fe2887ce568f7d6bf2c236" Dec 09 15:58:54 crc kubenswrapper[4894]: I1209 15:58:54.924197 4894 scope.go:117] "RemoveContainer" containerID="e8ff58696a142afbc97141c68c1505f7807ba4c434f4cc62d88c65a3afab76e5" Dec 09 15:58:54 crc kubenswrapper[4894]: I1209 15:58:54.962342 4894 scope.go:117] "RemoveContainer" containerID="fb9f558d8fc0848c266f1b70acc7ad89c12ec4368ceb418c4b5c2a60933c83ce" Dec 09 15:58:54 crc kubenswrapper[4894]: I1209 15:58:54.987741 4894 scope.go:117] "RemoveContainer" containerID="aa04b13b268897f23a37a764d1e04f0bd0ecf5b97c3effe0d1495a6f528e6ea6" Dec 09 15:58:55 crc kubenswrapper[4894]: I1209 15:58:55.006602 4894 scope.go:117] "RemoveContainer" containerID="da5a8a897a96863c3ab11ba7f22b4a58b681672b7e026cd0dc87fd9996e2d966" Dec 09 15:58:55 crc kubenswrapper[4894]: I1209 15:58:55.028991 4894 scope.go:117] "RemoveContainer" containerID="763c224aa9948df6b7bf361a9d72c078324d66c43ccdf343d608de6643c86769" Dec 09 15:58:55 crc kubenswrapper[4894]: I1209 15:58:55.049079 4894 scope.go:117] "RemoveContainer" containerID="a441a508256e662aad253db351dca851a1e46457e2b20ce5554b322ae16ffadc" Dec 09 15:58:55 crc kubenswrapper[4894]: I1209 15:58:55.090042 4894 scope.go:117] "RemoveContainer" containerID="9356d569802254937d2a4dccdd6757c5c7d3bf6a4595e2f8335a3ba17b1703c0" Dec 09 15:58:55 crc kubenswrapper[4894]: I1209 15:58:55.107965 4894 scope.go:117] "RemoveContainer" containerID="e58e180c7e9f42cce467df5cf3d928383cd93d1b7626a554d1b4b50b83382cf7" Dec 09 15:58:57 crc kubenswrapper[4894]: I1209 15:58:57.106361 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 15:58:57 crc kubenswrapper[4894]: E1209 15:58:57.106835 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 15:58:58 crc kubenswrapper[4894]: I1209 15:58:58.055172 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-npqvf"] Dec 09 15:58:58 crc kubenswrapper[4894]: I1209 15:58:58.065302 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-npqvf"] Dec 09 15:58:58 crc kubenswrapper[4894]: I1209 15:58:58.117631 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87072774-8453-49d2-a355-3388179d06dc" path="/var/lib/kubelet/pods/87072774-8453-49d2-a355-3388179d06dc/volumes" Dec 09 15:59:10 crc kubenswrapper[4894]: I1209 15:59:10.106555 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 15:59:10 crc kubenswrapper[4894]: E1209 15:59:10.107466 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 15:59:13 crc kubenswrapper[4894]: I1209 15:59:13.747459 4894 generic.go:334] "Generic (PLEG): container finished" podID="db9fb459-0a0a-4a5f-8510-12adae609c7f" containerID="55c85647833560d5e7e7e230a2b12b9f0ed5bce0cb39436f7e3ea135e2061620" exitCode=0 Dec 09 15:59:13 crc kubenswrapper[4894]: I1209 15:59:13.747566 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p" event={"ID":"db9fb459-0a0a-4a5f-8510-12adae609c7f","Type":"ContainerDied","Data":"55c85647833560d5e7e7e230a2b12b9f0ed5bce0cb39436f7e3ea135e2061620"} Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.234938 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.332967 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db9fb459-0a0a-4a5f-8510-12adae609c7f-ssh-key\") pod \"db9fb459-0a0a-4a5f-8510-12adae609c7f\" (UID: \"db9fb459-0a0a-4a5f-8510-12adae609c7f\") " Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.333137 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4p48g\" (UniqueName: \"kubernetes.io/projected/db9fb459-0a0a-4a5f-8510-12adae609c7f-kube-api-access-4p48g\") pod \"db9fb459-0a0a-4a5f-8510-12adae609c7f\" (UID: \"db9fb459-0a0a-4a5f-8510-12adae609c7f\") " Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.333412 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db9fb459-0a0a-4a5f-8510-12adae609c7f-inventory\") pod \"db9fb459-0a0a-4a5f-8510-12adae609c7f\" (UID: \"db9fb459-0a0a-4a5f-8510-12adae609c7f\") " Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.340761 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db9fb459-0a0a-4a5f-8510-12adae609c7f-kube-api-access-4p48g" (OuterVolumeSpecName: "kube-api-access-4p48g") pod "db9fb459-0a0a-4a5f-8510-12adae609c7f" (UID: "db9fb459-0a0a-4a5f-8510-12adae609c7f"). InnerVolumeSpecName "kube-api-access-4p48g". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.359932 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db9fb459-0a0a-4a5f-8510-12adae609c7f-inventory" (OuterVolumeSpecName: "inventory") pod "db9fb459-0a0a-4a5f-8510-12adae609c7f" (UID: "db9fb459-0a0a-4a5f-8510-12adae609c7f"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.363831 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db9fb459-0a0a-4a5f-8510-12adae609c7f-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "db9fb459-0a0a-4a5f-8510-12adae609c7f" (UID: "db9fb459-0a0a-4a5f-8510-12adae609c7f"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.436172 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/db9fb459-0a0a-4a5f-8510-12adae609c7f-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.436210 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4p48g\" (UniqueName: \"kubernetes.io/projected/db9fb459-0a0a-4a5f-8510-12adae609c7f-kube-api-access-4p48g\") on node \"crc\" DevicePath \"\"" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.436227 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/db9fb459-0a0a-4a5f-8510-12adae609c7f-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.774303 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p" event={"ID":"db9fb459-0a0a-4a5f-8510-12adae609c7f","Type":"ContainerDied","Data":"fab29074615c1463257fdb43179229f2cc8f508ffe56b39ef48e35c107e9ae90"} Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.774346 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.774364 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fab29074615c1463257fdb43179229f2cc8f508ffe56b39ef48e35c107e9ae90" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.865490 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx"] Dec 09 15:59:15 crc kubenswrapper[4894]: E1209 15:59:15.865906 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48095069-073b-4ea6-9608-2723fae94f24" containerName="registry-server" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.865924 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="48095069-073b-4ea6-9608-2723fae94f24" containerName="registry-server" Dec 09 15:59:15 crc kubenswrapper[4894]: E1209 15:59:15.865939 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db9fb459-0a0a-4a5f-8510-12adae609c7f" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.865946 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="db9fb459-0a0a-4a5f-8510-12adae609c7f" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 09 15:59:15 crc kubenswrapper[4894]: E1209 15:59:15.865975 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48095069-073b-4ea6-9608-2723fae94f24" containerName="extract-utilities" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.865981 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="48095069-073b-4ea6-9608-2723fae94f24" containerName="extract-utilities" Dec 09 15:59:15 crc kubenswrapper[4894]: E1209 15:59:15.865992 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="48095069-073b-4ea6-9608-2723fae94f24" containerName="extract-content" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.865999 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="48095069-073b-4ea6-9608-2723fae94f24" containerName="extract-content" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.866161 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="48095069-073b-4ea6-9608-2723fae94f24" containerName="registry-server" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.866183 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="db9fb459-0a0a-4a5f-8510-12adae609c7f" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.866772 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.868872 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.869261 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.869308 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.870037 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.885827 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx"] Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.946909 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3ac5cab-cda9-48f5-b042-3cc149a120d5-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx\" (UID: \"f3ac5cab-cda9-48f5-b042-3cc149a120d5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.946978 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3ac5cab-cda9-48f5-b042-3cc149a120d5-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx\" (UID: \"f3ac5cab-cda9-48f5-b042-3cc149a120d5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx" Dec 09 15:59:15 crc kubenswrapper[4894]: I1209 15:59:15.947124 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rx8z9\" (UniqueName: \"kubernetes.io/projected/f3ac5cab-cda9-48f5-b042-3cc149a120d5-kube-api-access-rx8z9\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx\" (UID: \"f3ac5cab-cda9-48f5-b042-3cc149a120d5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx" Dec 09 15:59:16 crc kubenswrapper[4894]: I1209 15:59:16.050323 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rx8z9\" (UniqueName: \"kubernetes.io/projected/f3ac5cab-cda9-48f5-b042-3cc149a120d5-kube-api-access-rx8z9\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx\" (UID: \"f3ac5cab-cda9-48f5-b042-3cc149a120d5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx" Dec 09 15:59:16 crc kubenswrapper[4894]: I1209 15:59:16.050407 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3ac5cab-cda9-48f5-b042-3cc149a120d5-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx\" (UID: \"f3ac5cab-cda9-48f5-b042-3cc149a120d5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx" Dec 09 15:59:16 crc kubenswrapper[4894]: I1209 15:59:16.050459 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3ac5cab-cda9-48f5-b042-3cc149a120d5-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx\" (UID: \"f3ac5cab-cda9-48f5-b042-3cc149a120d5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx" Dec 09 15:59:16 crc kubenswrapper[4894]: I1209 15:59:16.058441 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3ac5cab-cda9-48f5-b042-3cc149a120d5-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx\" (UID: \"f3ac5cab-cda9-48f5-b042-3cc149a120d5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx" Dec 09 15:59:16 crc kubenswrapper[4894]: I1209 15:59:16.058628 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3ac5cab-cda9-48f5-b042-3cc149a120d5-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx\" (UID: \"f3ac5cab-cda9-48f5-b042-3cc149a120d5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx" Dec 09 15:59:16 crc kubenswrapper[4894]: I1209 15:59:16.070947 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rx8z9\" (UniqueName: \"kubernetes.io/projected/f3ac5cab-cda9-48f5-b042-3cc149a120d5-kube-api-access-rx8z9\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx\" (UID: \"f3ac5cab-cda9-48f5-b042-3cc149a120d5\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx" Dec 09 15:59:16 crc kubenswrapper[4894]: I1209 15:59:16.190439 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx" Dec 09 15:59:16 crc kubenswrapper[4894]: I1209 15:59:16.553957 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx"] Dec 09 15:59:16 crc kubenswrapper[4894]: I1209 15:59:16.786733 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx" event={"ID":"f3ac5cab-cda9-48f5-b042-3cc149a120d5","Type":"ContainerStarted","Data":"2befe7d0baad49a498915ee3cb94d87b8a6fed9513730b9cf559480fb6d6ff0e"} Dec 09 15:59:17 crc kubenswrapper[4894]: I1209 15:59:17.798853 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx" event={"ID":"f3ac5cab-cda9-48f5-b042-3cc149a120d5","Type":"ContainerStarted","Data":"ee1242f966f9341754416dc1c1559f4b62f133d92681eccbec0c9269d274d8e5"} Dec 09 15:59:17 crc kubenswrapper[4894]: I1209 15:59:17.828377 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx" podStartSLOduration=2.033915833 podStartE2EDuration="2.828357822s" podCreationTimestamp="2025-12-09 15:59:15 +0000 UTC" firstStartedPulling="2025-12-09 15:59:16.561069822 +0000 UTC m=+1650.880280491" lastFinishedPulling="2025-12-09 15:59:17.355511771 +0000 UTC m=+1651.674722480" observedRunningTime="2025-12-09 15:59:17.818965749 +0000 UTC m=+1652.138176418" watchObservedRunningTime="2025-12-09 15:59:17.828357822 +0000 UTC m=+1652.147568491" Dec 09 15:59:22 crc kubenswrapper[4894]: I1209 15:59:22.107005 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 15:59:22 crc kubenswrapper[4894]: E1209 15:59:22.107857 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 15:59:22 crc kubenswrapper[4894]: I1209 15:59:22.849990 4894 generic.go:334] "Generic (PLEG): container finished" podID="f3ac5cab-cda9-48f5-b042-3cc149a120d5" containerID="ee1242f966f9341754416dc1c1559f4b62f133d92681eccbec0c9269d274d8e5" exitCode=0 Dec 09 15:59:22 crc kubenswrapper[4894]: I1209 15:59:22.850038 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx" event={"ID":"f3ac5cab-cda9-48f5-b042-3cc149a120d5","Type":"ContainerDied","Data":"ee1242f966f9341754416dc1c1559f4b62f133d92681eccbec0c9269d274d8e5"} Dec 09 15:59:24 crc kubenswrapper[4894]: I1209 15:59:24.276707 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx" Dec 09 15:59:24 crc kubenswrapper[4894]: I1209 15:59:24.301476 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3ac5cab-cda9-48f5-b042-3cc149a120d5-ssh-key\") pod \"f3ac5cab-cda9-48f5-b042-3cc149a120d5\" (UID: \"f3ac5cab-cda9-48f5-b042-3cc149a120d5\") " Dec 09 15:59:24 crc kubenswrapper[4894]: I1209 15:59:24.301554 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rx8z9\" (UniqueName: \"kubernetes.io/projected/f3ac5cab-cda9-48f5-b042-3cc149a120d5-kube-api-access-rx8z9\") pod \"f3ac5cab-cda9-48f5-b042-3cc149a120d5\" (UID: \"f3ac5cab-cda9-48f5-b042-3cc149a120d5\") " Dec 09 15:59:24 crc kubenswrapper[4894]: I1209 15:59:24.301619 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3ac5cab-cda9-48f5-b042-3cc149a120d5-inventory\") pod \"f3ac5cab-cda9-48f5-b042-3cc149a120d5\" (UID: \"f3ac5cab-cda9-48f5-b042-3cc149a120d5\") " Dec 09 15:59:24 crc kubenswrapper[4894]: I1209 15:59:24.309075 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3ac5cab-cda9-48f5-b042-3cc149a120d5-kube-api-access-rx8z9" (OuterVolumeSpecName: "kube-api-access-rx8z9") pod "f3ac5cab-cda9-48f5-b042-3cc149a120d5" (UID: "f3ac5cab-cda9-48f5-b042-3cc149a120d5"). InnerVolumeSpecName "kube-api-access-rx8z9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 15:59:24 crc kubenswrapper[4894]: I1209 15:59:24.330372 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3ac5cab-cda9-48f5-b042-3cc149a120d5-inventory" (OuterVolumeSpecName: "inventory") pod "f3ac5cab-cda9-48f5-b042-3cc149a120d5" (UID: "f3ac5cab-cda9-48f5-b042-3cc149a120d5"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:59:24 crc kubenswrapper[4894]: I1209 15:59:24.338255 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3ac5cab-cda9-48f5-b042-3cc149a120d5-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f3ac5cab-cda9-48f5-b042-3cc149a120d5" (UID: "f3ac5cab-cda9-48f5-b042-3cc149a120d5"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 15:59:24 crc kubenswrapper[4894]: I1209 15:59:24.403981 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3ac5cab-cda9-48f5-b042-3cc149a120d5-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 15:59:24 crc kubenswrapper[4894]: I1209 15:59:24.404028 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rx8z9\" (UniqueName: \"kubernetes.io/projected/f3ac5cab-cda9-48f5-b042-3cc149a120d5-kube-api-access-rx8z9\") on node \"crc\" DevicePath \"\"" Dec 09 15:59:24 crc kubenswrapper[4894]: I1209 15:59:24.404043 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3ac5cab-cda9-48f5-b042-3cc149a120d5-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 15:59:24 crc kubenswrapper[4894]: I1209 15:59:24.868059 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx" event={"ID":"f3ac5cab-cda9-48f5-b042-3cc149a120d5","Type":"ContainerDied","Data":"2befe7d0baad49a498915ee3cb94d87b8a6fed9513730b9cf559480fb6d6ff0e"} Dec 09 15:59:24 crc kubenswrapper[4894]: I1209 15:59:24.868099 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx" Dec 09 15:59:24 crc kubenswrapper[4894]: I1209 15:59:24.868112 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2befe7d0baad49a498915ee3cb94d87b8a6fed9513730b9cf559480fb6d6ff0e" Dec 09 15:59:24 crc kubenswrapper[4894]: I1209 15:59:24.961141 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f"] Dec 09 15:59:24 crc kubenswrapper[4894]: E1209 15:59:24.961560 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3ac5cab-cda9-48f5-b042-3cc149a120d5" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 09 15:59:24 crc kubenswrapper[4894]: I1209 15:59:24.961583 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3ac5cab-cda9-48f5-b042-3cc149a120d5" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 09 15:59:24 crc kubenswrapper[4894]: I1209 15:59:24.961847 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3ac5cab-cda9-48f5-b042-3cc149a120d5" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 09 15:59:24 crc kubenswrapper[4894]: I1209 15:59:24.962511 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f" Dec 09 15:59:24 crc kubenswrapper[4894]: I1209 15:59:24.966278 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 15:59:24 crc kubenswrapper[4894]: I1209 15:59:24.966428 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 15:59:24 crc kubenswrapper[4894]: I1209 15:59:24.967296 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 15:59:24 crc kubenswrapper[4894]: I1209 15:59:24.973945 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f"] Dec 09 15:59:24 crc kubenswrapper[4894]: I1209 15:59:24.976293 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 15:59:25 crc kubenswrapper[4894]: I1209 15:59:25.015110 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e6b8027-8cc2-460e-8749-3ceebeee52ad-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qk49f\" (UID: \"9e6b8027-8cc2-460e-8749-3ceebeee52ad\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f" Dec 09 15:59:25 crc kubenswrapper[4894]: I1209 15:59:25.015152 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lcrpq\" (UniqueName: \"kubernetes.io/projected/9e6b8027-8cc2-460e-8749-3ceebeee52ad-kube-api-access-lcrpq\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qk49f\" (UID: \"9e6b8027-8cc2-460e-8749-3ceebeee52ad\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f" Dec 09 15:59:25 crc kubenswrapper[4894]: I1209 15:59:25.015190 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e6b8027-8cc2-460e-8749-3ceebeee52ad-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qk49f\" (UID: \"9e6b8027-8cc2-460e-8749-3ceebeee52ad\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f" Dec 09 15:59:25 crc kubenswrapper[4894]: I1209 15:59:25.116264 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e6b8027-8cc2-460e-8749-3ceebeee52ad-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qk49f\" (UID: \"9e6b8027-8cc2-460e-8749-3ceebeee52ad\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f" Dec 09 15:59:25 crc kubenswrapper[4894]: I1209 15:59:25.116321 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lcrpq\" (UniqueName: \"kubernetes.io/projected/9e6b8027-8cc2-460e-8749-3ceebeee52ad-kube-api-access-lcrpq\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qk49f\" (UID: \"9e6b8027-8cc2-460e-8749-3ceebeee52ad\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f" Dec 09 15:59:25 crc kubenswrapper[4894]: I1209 15:59:25.116374 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e6b8027-8cc2-460e-8749-3ceebeee52ad-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qk49f\" (UID: \"9e6b8027-8cc2-460e-8749-3ceebeee52ad\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f" Dec 09 15:59:25 crc kubenswrapper[4894]: I1209 15:59:25.122290 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e6b8027-8cc2-460e-8749-3ceebeee52ad-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qk49f\" (UID: \"9e6b8027-8cc2-460e-8749-3ceebeee52ad\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f" Dec 09 15:59:25 crc kubenswrapper[4894]: I1209 15:59:25.122336 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e6b8027-8cc2-460e-8749-3ceebeee52ad-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qk49f\" (UID: \"9e6b8027-8cc2-460e-8749-3ceebeee52ad\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f" Dec 09 15:59:25 crc kubenswrapper[4894]: I1209 15:59:25.139674 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lcrpq\" (UniqueName: \"kubernetes.io/projected/9e6b8027-8cc2-460e-8749-3ceebeee52ad-kube-api-access-lcrpq\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-qk49f\" (UID: \"9e6b8027-8cc2-460e-8749-3ceebeee52ad\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f" Dec 09 15:59:25 crc kubenswrapper[4894]: I1209 15:59:25.328064 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f" Dec 09 15:59:25 crc kubenswrapper[4894]: I1209 15:59:25.956596 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f"] Dec 09 15:59:26 crc kubenswrapper[4894]: I1209 15:59:26.889492 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f" event={"ID":"9e6b8027-8cc2-460e-8749-3ceebeee52ad","Type":"ContainerStarted","Data":"e3775888aaa02850b1ef13192ab58c184bf78ba07afa819acff12c67fd4138b3"} Dec 09 15:59:26 crc kubenswrapper[4894]: I1209 15:59:26.889792 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f" event={"ID":"9e6b8027-8cc2-460e-8749-3ceebeee52ad","Type":"ContainerStarted","Data":"ea73c73c1e999599dbfe78fa75a60bdfc8f4aa495bb79502927b9e6c3305c7ed"} Dec 09 15:59:33 crc kubenswrapper[4894]: I1209 15:59:33.045427 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f" podStartSLOduration=8.39751737 podStartE2EDuration="9.045391583s" podCreationTimestamp="2025-12-09 15:59:24 +0000 UTC" firstStartedPulling="2025-12-09 15:59:25.957720356 +0000 UTC m=+1660.276931025" lastFinishedPulling="2025-12-09 15:59:26.605594569 +0000 UTC m=+1660.924805238" observedRunningTime="2025-12-09 15:59:26.907120287 +0000 UTC m=+1661.226330976" watchObservedRunningTime="2025-12-09 15:59:33.045391583 +0000 UTC m=+1667.364602292" Dec 09 15:59:33 crc kubenswrapper[4894]: I1209 15:59:33.064861 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-lfx27"] Dec 09 15:59:33 crc kubenswrapper[4894]: I1209 15:59:33.081718 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-lfx27"] Dec 09 15:59:34 crc kubenswrapper[4894]: I1209 15:59:34.106801 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 15:59:34 crc kubenswrapper[4894]: E1209 15:59:34.107313 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 15:59:34 crc kubenswrapper[4894]: I1209 15:59:34.121693 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b48430a2-ef1d-4747-97a4-4b6933fcea4f" path="/var/lib/kubelet/pods/b48430a2-ef1d-4747-97a4-4b6933fcea4f/volumes" Dec 09 15:59:40 crc kubenswrapper[4894]: I1209 15:59:40.032823 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-jr2lq"] Dec 09 15:59:40 crc kubenswrapper[4894]: I1209 15:59:40.044377 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-jr2lq"] Dec 09 15:59:40 crc kubenswrapper[4894]: I1209 15:59:40.120492 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7d3e1277-28d7-4aa1-9121-75d859ae4688" path="/var/lib/kubelet/pods/7d3e1277-28d7-4aa1-9121-75d859ae4688/volumes" Dec 09 15:59:43 crc kubenswrapper[4894]: I1209 15:59:43.045632 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-rl92n"] Dec 09 15:59:43 crc kubenswrapper[4894]: I1209 15:59:43.065491 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-rl92n"] Dec 09 15:59:44 crc kubenswrapper[4894]: I1209 15:59:44.138098 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="febdd308-ca10-4ba3-a34c-b000517eaeeb" path="/var/lib/kubelet/pods/febdd308-ca10-4ba3-a34c-b000517eaeeb/volumes" Dec 09 15:59:47 crc kubenswrapper[4894]: I1209 15:59:47.106949 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 15:59:47 crc kubenswrapper[4894]: E1209 15:59:47.110133 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 15:59:53 crc kubenswrapper[4894]: I1209 15:59:53.044432 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-s4wr5"] Dec 09 15:59:53 crc kubenswrapper[4894]: I1209 15:59:53.058156 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-s4wr5"] Dec 09 15:59:54 crc kubenswrapper[4894]: I1209 15:59:54.117602 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d29fb637-ba21-4cf9-ae4d-cd9720dd5af4" path="/var/lib/kubelet/pods/d29fb637-ba21-4cf9-ae4d-cd9720dd5af4/volumes" Dec 09 15:59:55 crc kubenswrapper[4894]: I1209 15:59:55.035669 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-nbnz8"] Dec 09 15:59:55 crc kubenswrapper[4894]: I1209 15:59:55.046505 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-nbnz8"] Dec 09 15:59:55 crc kubenswrapper[4894]: I1209 15:59:55.395327 4894 scope.go:117] "RemoveContainer" containerID="b8a1f501c7d2757a2ff9a88f8716626e02ee6c30fae871162ab553918ea821dc" Dec 09 15:59:55 crc kubenswrapper[4894]: I1209 15:59:55.449825 4894 scope.go:117] "RemoveContainer" containerID="25bc8616e520df5b88b689fa0aa217ecd8bd11c7b2224dc7ce70a2f0697d7388" Dec 09 15:59:55 crc kubenswrapper[4894]: I1209 15:59:55.512868 4894 scope.go:117] "RemoveContainer" containerID="0abfc9e8d74dfb18b88220cbd3c45dca30fc929eacadd325f0ba4baa46b1b3e9" Dec 09 15:59:55 crc kubenswrapper[4894]: I1209 15:59:55.559400 4894 scope.go:117] "RemoveContainer" containerID="bf7b9361fff26dd243ab8d42e89b0ce1e58793f86d9fa6fd7519b1a4d80667d7" Dec 09 15:59:55 crc kubenswrapper[4894]: I1209 15:59:55.623518 4894 scope.go:117] "RemoveContainer" containerID="7a1928ef8d06e25be6e0fab76f6d003eaa7e0f5678729dc4598fa1ebcfbe1d67" Dec 09 15:59:56 crc kubenswrapper[4894]: I1209 15:59:56.124501 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ce9f3ab9-097d-4d0d-8082-08019bce834c" path="/var/lib/kubelet/pods/ce9f3ab9-097d-4d0d-8082-08019bce834c/volumes" Dec 09 15:59:59 crc kubenswrapper[4894]: I1209 15:59:59.107555 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 15:59:59 crc kubenswrapper[4894]: E1209 15:59:59.108892 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:00:00 crc kubenswrapper[4894]: I1209 16:00:00.161861 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz"] Dec 09 16:00:00 crc kubenswrapper[4894]: I1209 16:00:00.164069 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz" Dec 09 16:00:00 crc kubenswrapper[4894]: I1209 16:00:00.166274 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 16:00:00 crc kubenswrapper[4894]: I1209 16:00:00.166978 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 16:00:00 crc kubenswrapper[4894]: I1209 16:00:00.178270 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz"] Dec 09 16:00:00 crc kubenswrapper[4894]: I1209 16:00:00.282414 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/acd00528-8721-45e8-b085-c8dab62fbc7f-config-volume\") pod \"collect-profiles-29421600-t6btz\" (UID: \"acd00528-8721-45e8-b085-c8dab62fbc7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz" Dec 09 16:00:00 crc kubenswrapper[4894]: I1209 16:00:00.282495 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/acd00528-8721-45e8-b085-c8dab62fbc7f-secret-volume\") pod \"collect-profiles-29421600-t6btz\" (UID: \"acd00528-8721-45e8-b085-c8dab62fbc7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz" Dec 09 16:00:00 crc kubenswrapper[4894]: I1209 16:00:00.282539 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w8gcx\" (UniqueName: \"kubernetes.io/projected/acd00528-8721-45e8-b085-c8dab62fbc7f-kube-api-access-w8gcx\") pod \"collect-profiles-29421600-t6btz\" (UID: \"acd00528-8721-45e8-b085-c8dab62fbc7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz" Dec 09 16:00:00 crc kubenswrapper[4894]: I1209 16:00:00.385675 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/acd00528-8721-45e8-b085-c8dab62fbc7f-config-volume\") pod \"collect-profiles-29421600-t6btz\" (UID: \"acd00528-8721-45e8-b085-c8dab62fbc7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz" Dec 09 16:00:00 crc kubenswrapper[4894]: I1209 16:00:00.385771 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/acd00528-8721-45e8-b085-c8dab62fbc7f-secret-volume\") pod \"collect-profiles-29421600-t6btz\" (UID: \"acd00528-8721-45e8-b085-c8dab62fbc7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz" Dec 09 16:00:00 crc kubenswrapper[4894]: I1209 16:00:00.385833 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w8gcx\" (UniqueName: \"kubernetes.io/projected/acd00528-8721-45e8-b085-c8dab62fbc7f-kube-api-access-w8gcx\") pod \"collect-profiles-29421600-t6btz\" (UID: \"acd00528-8721-45e8-b085-c8dab62fbc7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz" Dec 09 16:00:00 crc kubenswrapper[4894]: I1209 16:00:00.387388 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/acd00528-8721-45e8-b085-c8dab62fbc7f-config-volume\") pod \"collect-profiles-29421600-t6btz\" (UID: \"acd00528-8721-45e8-b085-c8dab62fbc7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz" Dec 09 16:00:00 crc kubenswrapper[4894]: I1209 16:00:00.395970 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/acd00528-8721-45e8-b085-c8dab62fbc7f-secret-volume\") pod \"collect-profiles-29421600-t6btz\" (UID: \"acd00528-8721-45e8-b085-c8dab62fbc7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz" Dec 09 16:00:00 crc kubenswrapper[4894]: I1209 16:00:00.409769 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w8gcx\" (UniqueName: \"kubernetes.io/projected/acd00528-8721-45e8-b085-c8dab62fbc7f-kube-api-access-w8gcx\") pod \"collect-profiles-29421600-t6btz\" (UID: \"acd00528-8721-45e8-b085-c8dab62fbc7f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz" Dec 09 16:00:00 crc kubenswrapper[4894]: I1209 16:00:00.501424 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz" Dec 09 16:00:00 crc kubenswrapper[4894]: I1209 16:00:00.964332 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz"] Dec 09 16:00:01 crc kubenswrapper[4894]: I1209 16:00:01.240894 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz" event={"ID":"acd00528-8721-45e8-b085-c8dab62fbc7f","Type":"ContainerStarted","Data":"8529f745f32501dce0e99d50fb124aa049cd3448ed68dd213d19bec97dcd1c85"} Dec 09 16:00:01 crc kubenswrapper[4894]: I1209 16:00:01.241243 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz" event={"ID":"acd00528-8721-45e8-b085-c8dab62fbc7f","Type":"ContainerStarted","Data":"66613f9816397cb86bc5c20c756f9659b7c113c4bcdbc5a761445795929df6ff"} Dec 09 16:00:01 crc kubenswrapper[4894]: I1209 16:00:01.270082 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz" podStartSLOduration=1.270064276 podStartE2EDuration="1.270064276s" podCreationTimestamp="2025-12-09 16:00:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:00:01.2631186 +0000 UTC m=+1695.582329279" watchObservedRunningTime="2025-12-09 16:00:01.270064276 +0000 UTC m=+1695.589274945" Dec 09 16:00:02 crc kubenswrapper[4894]: I1209 16:00:02.249247 4894 generic.go:334] "Generic (PLEG): container finished" podID="acd00528-8721-45e8-b085-c8dab62fbc7f" containerID="8529f745f32501dce0e99d50fb124aa049cd3448ed68dd213d19bec97dcd1c85" exitCode=0 Dec 09 16:00:02 crc kubenswrapper[4894]: I1209 16:00:02.249312 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz" event={"ID":"acd00528-8721-45e8-b085-c8dab62fbc7f","Type":"ContainerDied","Data":"8529f745f32501dce0e99d50fb124aa049cd3448ed68dd213d19bec97dcd1c85"} Dec 09 16:00:03 crc kubenswrapper[4894]: I1209 16:00:03.594775 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz" Dec 09 16:00:03 crc kubenswrapper[4894]: I1209 16:00:03.744396 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/acd00528-8721-45e8-b085-c8dab62fbc7f-config-volume\") pod \"acd00528-8721-45e8-b085-c8dab62fbc7f\" (UID: \"acd00528-8721-45e8-b085-c8dab62fbc7f\") " Dec 09 16:00:03 crc kubenswrapper[4894]: I1209 16:00:03.744905 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/acd00528-8721-45e8-b085-c8dab62fbc7f-secret-volume\") pod \"acd00528-8721-45e8-b085-c8dab62fbc7f\" (UID: \"acd00528-8721-45e8-b085-c8dab62fbc7f\") " Dec 09 16:00:03 crc kubenswrapper[4894]: I1209 16:00:03.745017 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w8gcx\" (UniqueName: \"kubernetes.io/projected/acd00528-8721-45e8-b085-c8dab62fbc7f-kube-api-access-w8gcx\") pod \"acd00528-8721-45e8-b085-c8dab62fbc7f\" (UID: \"acd00528-8721-45e8-b085-c8dab62fbc7f\") " Dec 09 16:00:03 crc kubenswrapper[4894]: I1209 16:00:03.746983 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/acd00528-8721-45e8-b085-c8dab62fbc7f-config-volume" (OuterVolumeSpecName: "config-volume") pod "acd00528-8721-45e8-b085-c8dab62fbc7f" (UID: "acd00528-8721-45e8-b085-c8dab62fbc7f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:00:03 crc kubenswrapper[4894]: I1209 16:00:03.754353 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/acd00528-8721-45e8-b085-c8dab62fbc7f-kube-api-access-w8gcx" (OuterVolumeSpecName: "kube-api-access-w8gcx") pod "acd00528-8721-45e8-b085-c8dab62fbc7f" (UID: "acd00528-8721-45e8-b085-c8dab62fbc7f"). InnerVolumeSpecName "kube-api-access-w8gcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:00:03 crc kubenswrapper[4894]: I1209 16:00:03.754499 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/acd00528-8721-45e8-b085-c8dab62fbc7f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "acd00528-8721-45e8-b085-c8dab62fbc7f" (UID: "acd00528-8721-45e8-b085-c8dab62fbc7f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:00:03 crc kubenswrapper[4894]: I1209 16:00:03.847558 4894 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/acd00528-8721-45e8-b085-c8dab62fbc7f-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 16:00:03 crc kubenswrapper[4894]: I1209 16:00:03.847600 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w8gcx\" (UniqueName: \"kubernetes.io/projected/acd00528-8721-45e8-b085-c8dab62fbc7f-kube-api-access-w8gcx\") on node \"crc\" DevicePath \"\"" Dec 09 16:00:03 crc kubenswrapper[4894]: I1209 16:00:03.847609 4894 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/acd00528-8721-45e8-b085-c8dab62fbc7f-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 16:00:04 crc kubenswrapper[4894]: I1209 16:00:04.269150 4894 generic.go:334] "Generic (PLEG): container finished" podID="9e6b8027-8cc2-460e-8749-3ceebeee52ad" containerID="e3775888aaa02850b1ef13192ab58c184bf78ba07afa819acff12c67fd4138b3" exitCode=0 Dec 09 16:00:04 crc kubenswrapper[4894]: I1209 16:00:04.269238 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f" event={"ID":"9e6b8027-8cc2-460e-8749-3ceebeee52ad","Type":"ContainerDied","Data":"e3775888aaa02850b1ef13192ab58c184bf78ba07afa819acff12c67fd4138b3"} Dec 09 16:00:04 crc kubenswrapper[4894]: I1209 16:00:04.272763 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz" event={"ID":"acd00528-8721-45e8-b085-c8dab62fbc7f","Type":"ContainerDied","Data":"66613f9816397cb86bc5c20c756f9659b7c113c4bcdbc5a761445795929df6ff"} Dec 09 16:00:04 crc kubenswrapper[4894]: I1209 16:00:04.272826 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="66613f9816397cb86bc5c20c756f9659b7c113c4bcdbc5a761445795929df6ff" Dec 09 16:00:04 crc kubenswrapper[4894]: I1209 16:00:04.272851 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz" Dec 09 16:00:05 crc kubenswrapper[4894]: I1209 16:00:05.739613 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f" Dec 09 16:00:05 crc kubenswrapper[4894]: I1209 16:00:05.888777 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lcrpq\" (UniqueName: \"kubernetes.io/projected/9e6b8027-8cc2-460e-8749-3ceebeee52ad-kube-api-access-lcrpq\") pod \"9e6b8027-8cc2-460e-8749-3ceebeee52ad\" (UID: \"9e6b8027-8cc2-460e-8749-3ceebeee52ad\") " Dec 09 16:00:05 crc kubenswrapper[4894]: I1209 16:00:05.888839 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e6b8027-8cc2-460e-8749-3ceebeee52ad-inventory\") pod \"9e6b8027-8cc2-460e-8749-3ceebeee52ad\" (UID: \"9e6b8027-8cc2-460e-8749-3ceebeee52ad\") " Dec 09 16:00:05 crc kubenswrapper[4894]: I1209 16:00:05.888953 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e6b8027-8cc2-460e-8749-3ceebeee52ad-ssh-key\") pod \"9e6b8027-8cc2-460e-8749-3ceebeee52ad\" (UID: \"9e6b8027-8cc2-460e-8749-3ceebeee52ad\") " Dec 09 16:00:05 crc kubenswrapper[4894]: I1209 16:00:05.896351 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9e6b8027-8cc2-460e-8749-3ceebeee52ad-kube-api-access-lcrpq" (OuterVolumeSpecName: "kube-api-access-lcrpq") pod "9e6b8027-8cc2-460e-8749-3ceebeee52ad" (UID: "9e6b8027-8cc2-460e-8749-3ceebeee52ad"). InnerVolumeSpecName "kube-api-access-lcrpq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:00:05 crc kubenswrapper[4894]: I1209 16:00:05.927395 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e6b8027-8cc2-460e-8749-3ceebeee52ad-inventory" (OuterVolumeSpecName: "inventory") pod "9e6b8027-8cc2-460e-8749-3ceebeee52ad" (UID: "9e6b8027-8cc2-460e-8749-3ceebeee52ad"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:00:05 crc kubenswrapper[4894]: I1209 16:00:05.936836 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9e6b8027-8cc2-460e-8749-3ceebeee52ad-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9e6b8027-8cc2-460e-8749-3ceebeee52ad" (UID: "9e6b8027-8cc2-460e-8749-3ceebeee52ad"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:00:05 crc kubenswrapper[4894]: I1209 16:00:05.991695 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lcrpq\" (UniqueName: \"kubernetes.io/projected/9e6b8027-8cc2-460e-8749-3ceebeee52ad-kube-api-access-lcrpq\") on node \"crc\" DevicePath \"\"" Dec 09 16:00:05 crc kubenswrapper[4894]: I1209 16:00:05.991751 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9e6b8027-8cc2-460e-8749-3ceebeee52ad-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 16:00:05 crc kubenswrapper[4894]: I1209 16:00:05.991767 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9e6b8027-8cc2-460e-8749-3ceebeee52ad-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.296735 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f" event={"ID":"9e6b8027-8cc2-460e-8749-3ceebeee52ad","Type":"ContainerDied","Data":"ea73c73c1e999599dbfe78fa75a60bdfc8f4aa495bb79502927b9e6c3305c7ed"} Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.296790 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ea73c73c1e999599dbfe78fa75a60bdfc8f4aa495bb79502927b9e6c3305c7ed" Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.296825 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f" Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.386291 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b"] Dec 09 16:00:06 crc kubenswrapper[4894]: E1209 16:00:06.386847 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="acd00528-8721-45e8-b085-c8dab62fbc7f" containerName="collect-profiles" Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.386870 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="acd00528-8721-45e8-b085-c8dab62fbc7f" containerName="collect-profiles" Dec 09 16:00:06 crc kubenswrapper[4894]: E1209 16:00:06.386924 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9e6b8027-8cc2-460e-8749-3ceebeee52ad" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.386938 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9e6b8027-8cc2-460e-8749-3ceebeee52ad" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.387228 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="acd00528-8721-45e8-b085-c8dab62fbc7f" containerName="collect-profiles" Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.387314 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9e6b8027-8cc2-460e-8749-3ceebeee52ad" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.388351 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b" Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.392501 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.393161 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.393518 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.393820 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.404951 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b"] Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.504289 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l5ftf\" (UniqueName: \"kubernetes.io/projected/b5ef8626-14e1-4fb2-a137-2bf35b64146d-kube-api-access-l5ftf\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b\" (UID: \"b5ef8626-14e1-4fb2-a137-2bf35b64146d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b" Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.504573 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b5ef8626-14e1-4fb2-a137-2bf35b64146d-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b\" (UID: \"b5ef8626-14e1-4fb2-a137-2bf35b64146d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b" Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.504621 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5ef8626-14e1-4fb2-a137-2bf35b64146d-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b\" (UID: \"b5ef8626-14e1-4fb2-a137-2bf35b64146d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b" Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.606753 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l5ftf\" (UniqueName: \"kubernetes.io/projected/b5ef8626-14e1-4fb2-a137-2bf35b64146d-kube-api-access-l5ftf\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b\" (UID: \"b5ef8626-14e1-4fb2-a137-2bf35b64146d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b" Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.607138 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b5ef8626-14e1-4fb2-a137-2bf35b64146d-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b\" (UID: \"b5ef8626-14e1-4fb2-a137-2bf35b64146d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b" Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.607223 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5ef8626-14e1-4fb2-a137-2bf35b64146d-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b\" (UID: \"b5ef8626-14e1-4fb2-a137-2bf35b64146d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b" Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.615151 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b5ef8626-14e1-4fb2-a137-2bf35b64146d-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b\" (UID: \"b5ef8626-14e1-4fb2-a137-2bf35b64146d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b" Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.617309 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5ef8626-14e1-4fb2-a137-2bf35b64146d-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b\" (UID: \"b5ef8626-14e1-4fb2-a137-2bf35b64146d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b" Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.638690 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l5ftf\" (UniqueName: \"kubernetes.io/projected/b5ef8626-14e1-4fb2-a137-2bf35b64146d-kube-api-access-l5ftf\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b\" (UID: \"b5ef8626-14e1-4fb2-a137-2bf35b64146d\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b" Dec 09 16:00:06 crc kubenswrapper[4894]: I1209 16:00:06.730428 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b" Dec 09 16:00:07 crc kubenswrapper[4894]: I1209 16:00:07.057767 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b"] Dec 09 16:00:07 crc kubenswrapper[4894]: I1209 16:00:07.308972 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b" event={"ID":"b5ef8626-14e1-4fb2-a137-2bf35b64146d","Type":"ContainerStarted","Data":"4bd56a52428b263a97dce053b52c158396ed18b48d13c25de3c1ba0064c8e30a"} Dec 09 16:00:08 crc kubenswrapper[4894]: I1209 16:00:08.319141 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b" event={"ID":"b5ef8626-14e1-4fb2-a137-2bf35b64146d","Type":"ContainerStarted","Data":"50066c981c62fafecfce155649674f54334eb95319b66113df81cb777e116108"} Dec 09 16:00:08 crc kubenswrapper[4894]: I1209 16:00:08.343612 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b" podStartSLOduration=1.7651404990000001 podStartE2EDuration="2.343596163s" podCreationTimestamp="2025-12-09 16:00:06 +0000 UTC" firstStartedPulling="2025-12-09 16:00:07.060715993 +0000 UTC m=+1701.379926662" lastFinishedPulling="2025-12-09 16:00:07.639171647 +0000 UTC m=+1701.958382326" observedRunningTime="2025-12-09 16:00:08.335701711 +0000 UTC m=+1702.654912400" watchObservedRunningTime="2025-12-09 16:00:08.343596163 +0000 UTC m=+1702.662806832" Dec 09 16:00:10 crc kubenswrapper[4894]: I1209 16:00:10.106036 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 16:00:10 crc kubenswrapper[4894]: E1209 16:00:10.106673 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:00:12 crc kubenswrapper[4894]: I1209 16:00:12.363464 4894 generic.go:334] "Generic (PLEG): container finished" podID="b5ef8626-14e1-4fb2-a137-2bf35b64146d" containerID="50066c981c62fafecfce155649674f54334eb95319b66113df81cb777e116108" exitCode=0 Dec 09 16:00:12 crc kubenswrapper[4894]: I1209 16:00:12.363680 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b" event={"ID":"b5ef8626-14e1-4fb2-a137-2bf35b64146d","Type":"ContainerDied","Data":"50066c981c62fafecfce155649674f54334eb95319b66113df81cb777e116108"} Dec 09 16:00:13 crc kubenswrapper[4894]: I1209 16:00:13.755424 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b" Dec 09 16:00:13 crc kubenswrapper[4894]: I1209 16:00:13.841435 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l5ftf\" (UniqueName: \"kubernetes.io/projected/b5ef8626-14e1-4fb2-a137-2bf35b64146d-kube-api-access-l5ftf\") pod \"b5ef8626-14e1-4fb2-a137-2bf35b64146d\" (UID: \"b5ef8626-14e1-4fb2-a137-2bf35b64146d\") " Dec 09 16:00:13 crc kubenswrapper[4894]: I1209 16:00:13.841554 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b5ef8626-14e1-4fb2-a137-2bf35b64146d-ssh-key\") pod \"b5ef8626-14e1-4fb2-a137-2bf35b64146d\" (UID: \"b5ef8626-14e1-4fb2-a137-2bf35b64146d\") " Dec 09 16:00:13 crc kubenswrapper[4894]: I1209 16:00:13.841746 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5ef8626-14e1-4fb2-a137-2bf35b64146d-inventory\") pod \"b5ef8626-14e1-4fb2-a137-2bf35b64146d\" (UID: \"b5ef8626-14e1-4fb2-a137-2bf35b64146d\") " Dec 09 16:00:13 crc kubenswrapper[4894]: I1209 16:00:13.848558 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b5ef8626-14e1-4fb2-a137-2bf35b64146d-kube-api-access-l5ftf" (OuterVolumeSpecName: "kube-api-access-l5ftf") pod "b5ef8626-14e1-4fb2-a137-2bf35b64146d" (UID: "b5ef8626-14e1-4fb2-a137-2bf35b64146d"). InnerVolumeSpecName "kube-api-access-l5ftf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:00:13 crc kubenswrapper[4894]: I1209 16:00:13.867897 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5ef8626-14e1-4fb2-a137-2bf35b64146d-inventory" (OuterVolumeSpecName: "inventory") pod "b5ef8626-14e1-4fb2-a137-2bf35b64146d" (UID: "b5ef8626-14e1-4fb2-a137-2bf35b64146d"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:00:13 crc kubenswrapper[4894]: I1209 16:00:13.868950 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b5ef8626-14e1-4fb2-a137-2bf35b64146d-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "b5ef8626-14e1-4fb2-a137-2bf35b64146d" (UID: "b5ef8626-14e1-4fb2-a137-2bf35b64146d"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:00:13 crc kubenswrapper[4894]: I1209 16:00:13.944542 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/b5ef8626-14e1-4fb2-a137-2bf35b64146d-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:00:13 crc kubenswrapper[4894]: I1209 16:00:13.944584 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/b5ef8626-14e1-4fb2-a137-2bf35b64146d-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 16:00:13 crc kubenswrapper[4894]: I1209 16:00:13.944597 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l5ftf\" (UniqueName: \"kubernetes.io/projected/b5ef8626-14e1-4fb2-a137-2bf35b64146d-kube-api-access-l5ftf\") on node \"crc\" DevicePath \"\"" Dec 09 16:00:14 crc kubenswrapper[4894]: I1209 16:00:14.383067 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b" event={"ID":"b5ef8626-14e1-4fb2-a137-2bf35b64146d","Type":"ContainerDied","Data":"4bd56a52428b263a97dce053b52c158396ed18b48d13c25de3c1ba0064c8e30a"} Dec 09 16:00:14 crc kubenswrapper[4894]: I1209 16:00:14.383444 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4bd56a52428b263a97dce053b52c158396ed18b48d13c25de3c1ba0064c8e30a" Dec 09 16:00:14 crc kubenswrapper[4894]: I1209 16:00:14.383161 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b" Dec 09 16:00:14 crc kubenswrapper[4894]: I1209 16:00:14.463725 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp"] Dec 09 16:00:14 crc kubenswrapper[4894]: E1209 16:00:14.464169 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b5ef8626-14e1-4fb2-a137-2bf35b64146d" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 09 16:00:14 crc kubenswrapper[4894]: I1209 16:00:14.464197 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b5ef8626-14e1-4fb2-a137-2bf35b64146d" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 09 16:00:14 crc kubenswrapper[4894]: I1209 16:00:14.464426 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b5ef8626-14e1-4fb2-a137-2bf35b64146d" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 09 16:00:14 crc kubenswrapper[4894]: I1209 16:00:14.465149 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp" Dec 09 16:00:14 crc kubenswrapper[4894]: I1209 16:00:14.468318 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 16:00:14 crc kubenswrapper[4894]: I1209 16:00:14.468717 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 16:00:14 crc kubenswrapper[4894]: I1209 16:00:14.469065 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 16:00:14 crc kubenswrapper[4894]: I1209 16:00:14.470909 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 16:00:14 crc kubenswrapper[4894]: I1209 16:00:14.482324 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp"] Dec 09 16:00:14 crc kubenswrapper[4894]: I1209 16:00:14.555853 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94a3b079-5c34-4ef3-9262-eb741ef7910b-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp\" (UID: \"94a3b079-5c34-4ef3-9262-eb741ef7910b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp" Dec 09 16:00:14 crc kubenswrapper[4894]: I1209 16:00:14.555945 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-84znz\" (UniqueName: \"kubernetes.io/projected/94a3b079-5c34-4ef3-9262-eb741ef7910b-kube-api-access-84znz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp\" (UID: \"94a3b079-5c34-4ef3-9262-eb741ef7910b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp" Dec 09 16:00:14 crc kubenswrapper[4894]: I1209 16:00:14.556243 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94a3b079-5c34-4ef3-9262-eb741ef7910b-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp\" (UID: \"94a3b079-5c34-4ef3-9262-eb741ef7910b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp" Dec 09 16:00:14 crc kubenswrapper[4894]: I1209 16:00:14.657957 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94a3b079-5c34-4ef3-9262-eb741ef7910b-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp\" (UID: \"94a3b079-5c34-4ef3-9262-eb741ef7910b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp" Dec 09 16:00:14 crc kubenswrapper[4894]: I1209 16:00:14.658039 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-84znz\" (UniqueName: \"kubernetes.io/projected/94a3b079-5c34-4ef3-9262-eb741ef7910b-kube-api-access-84znz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp\" (UID: \"94a3b079-5c34-4ef3-9262-eb741ef7910b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp" Dec 09 16:00:14 crc kubenswrapper[4894]: I1209 16:00:14.658135 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94a3b079-5c34-4ef3-9262-eb741ef7910b-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp\" (UID: \"94a3b079-5c34-4ef3-9262-eb741ef7910b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp" Dec 09 16:00:14 crc kubenswrapper[4894]: I1209 16:00:14.663201 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94a3b079-5c34-4ef3-9262-eb741ef7910b-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp\" (UID: \"94a3b079-5c34-4ef3-9262-eb741ef7910b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp" Dec 09 16:00:14 crc kubenswrapper[4894]: I1209 16:00:14.671332 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94a3b079-5c34-4ef3-9262-eb741ef7910b-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp\" (UID: \"94a3b079-5c34-4ef3-9262-eb741ef7910b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp" Dec 09 16:00:14 crc kubenswrapper[4894]: I1209 16:00:14.685216 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-84znz\" (UniqueName: \"kubernetes.io/projected/94a3b079-5c34-4ef3-9262-eb741ef7910b-kube-api-access-84znz\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp\" (UID: \"94a3b079-5c34-4ef3-9262-eb741ef7910b\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp" Dec 09 16:00:14 crc kubenswrapper[4894]: I1209 16:00:14.781953 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp" Dec 09 16:00:15 crc kubenswrapper[4894]: I1209 16:00:15.338124 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp"] Dec 09 16:00:15 crc kubenswrapper[4894]: I1209 16:00:15.391991 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp" event={"ID":"94a3b079-5c34-4ef3-9262-eb741ef7910b","Type":"ContainerStarted","Data":"188fc594cea51c37032d8464419192d1057841d3721f2c2500b8b57d12fb1d54"} Dec 09 16:00:17 crc kubenswrapper[4894]: I1209 16:00:17.411407 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp" event={"ID":"94a3b079-5c34-4ef3-9262-eb741ef7910b","Type":"ContainerStarted","Data":"ad356faf9f26c46787e5421cec638ad9dee018928d9909ef8762fe781d29ed10"} Dec 09 16:00:17 crc kubenswrapper[4894]: I1209 16:00:17.434863 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp" podStartSLOduration=2.511112756 podStartE2EDuration="3.434839504s" podCreationTimestamp="2025-12-09 16:00:14 +0000 UTC" firstStartedPulling="2025-12-09 16:00:15.347355302 +0000 UTC m=+1709.666565971" lastFinishedPulling="2025-12-09 16:00:16.27108203 +0000 UTC m=+1710.590292719" observedRunningTime="2025-12-09 16:00:17.428721389 +0000 UTC m=+1711.747932068" watchObservedRunningTime="2025-12-09 16:00:17.434839504 +0000 UTC m=+1711.754050183" Dec 09 16:00:22 crc kubenswrapper[4894]: I1209 16:00:22.107002 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 16:00:22 crc kubenswrapper[4894]: E1209 16:00:22.107735 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:00:33 crc kubenswrapper[4894]: I1209 16:00:33.107039 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 16:00:33 crc kubenswrapper[4894]: E1209 16:00:33.107925 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:00:43 crc kubenswrapper[4894]: I1209 16:00:43.075432 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-r8zsp"] Dec 09 16:00:43 crc kubenswrapper[4894]: I1209 16:00:43.084392 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-r8zsp"] Dec 09 16:00:44 crc kubenswrapper[4894]: I1209 16:00:44.120954 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea9963b9-ff7e-4d92-9cf1-7a5108a33b76" path="/var/lib/kubelet/pods/ea9963b9-ff7e-4d92-9cf1-7a5108a33b76/volumes" Dec 09 16:00:45 crc kubenswrapper[4894]: I1209 16:00:45.106703 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 16:00:45 crc kubenswrapper[4894]: E1209 16:00:45.107164 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:00:52 crc kubenswrapper[4894]: I1209 16:00:52.045811 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-5951-account-create-update-zl4pj"] Dec 09 16:00:52 crc kubenswrapper[4894]: I1209 16:00:52.060836 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-5951-account-create-update-zl4pj"] Dec 09 16:00:52 crc kubenswrapper[4894]: I1209 16:00:52.118614 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3baa21af-058e-41cc-832b-6e6b138a33d2" path="/var/lib/kubelet/pods/3baa21af-058e-41cc-832b-6e6b138a33d2/volumes" Dec 09 16:00:53 crc kubenswrapper[4894]: I1209 16:00:53.032248 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-vt68w"] Dec 09 16:00:53 crc kubenswrapper[4894]: I1209 16:00:53.042068 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-5x62k"] Dec 09 16:00:53 crc kubenswrapper[4894]: I1209 16:00:53.049984 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-vt68w"] Dec 09 16:00:53 crc kubenswrapper[4894]: I1209 16:00:53.057369 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-5x62k"] Dec 09 16:00:54 crc kubenswrapper[4894]: I1209 16:00:54.117688 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c" path="/var/lib/kubelet/pods/88a1beb7-7b8d-4887-b5d6-e59b9ffe9e3c/volumes" Dec 09 16:00:54 crc kubenswrapper[4894]: I1209 16:00:54.119473 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e66b850-9ac9-43ce-977a-a8f61c80a2da" path="/var/lib/kubelet/pods/8e66b850-9ac9-43ce-977a-a8f61c80a2da/volumes" Dec 09 16:00:55 crc kubenswrapper[4894]: I1209 16:00:55.756296 4894 scope.go:117] "RemoveContainer" containerID="19cca9c07ba48a5f43b63d2d9f6068a2048d5c05390daa973c00896a6e47eed3" Dec 09 16:00:55 crc kubenswrapper[4894]: I1209 16:00:55.800178 4894 scope.go:117] "RemoveContainer" containerID="21e1648dc1dccfed5f68bef50d540de6507f298f0a3b91b0fe33cf916a487e4f" Dec 09 16:00:55 crc kubenswrapper[4894]: I1209 16:00:55.829038 4894 scope.go:117] "RemoveContainer" containerID="f2e5fa88e421d711a8f091021fdd47f49dfbfe3a8d6e38495324d1020a7b71ea" Dec 09 16:00:55 crc kubenswrapper[4894]: I1209 16:00:55.885245 4894 scope.go:117] "RemoveContainer" containerID="c13c0b98297df1f9d64e1feab246d03958c398c3a45c527d5b51479615e8a04a" Dec 09 16:00:55 crc kubenswrapper[4894]: I1209 16:00:55.914702 4894 scope.go:117] "RemoveContainer" containerID="6b95ed24650ca49fa2e0af023c8af9711bcfa64f731d350e210e70cc20b97e37" Dec 09 16:00:56 crc kubenswrapper[4894]: I1209 16:00:56.114969 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 16:00:56 crc kubenswrapper[4894]: E1209 16:00:56.115813 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:01:00 crc kubenswrapper[4894]: I1209 16:01:00.150712 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-cron-29421601-2jlb4"] Dec 09 16:01:00 crc kubenswrapper[4894]: I1209 16:01:00.153028 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29421601-2jlb4" Dec 09 16:01:00 crc kubenswrapper[4894]: I1209 16:01:00.160333 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29421601-2jlb4"] Dec 09 16:01:00 crc kubenswrapper[4894]: I1209 16:01:00.165822 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629496e1-9961-4c1d-8f3c-9426f7ab0f72-combined-ca-bundle\") pod \"keystone-cron-29421601-2jlb4\" (UID: \"629496e1-9961-4c1d-8f3c-9426f7ab0f72\") " pod="openstack/keystone-cron-29421601-2jlb4" Dec 09 16:01:00 crc kubenswrapper[4894]: I1209 16:01:00.165892 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/629496e1-9961-4c1d-8f3c-9426f7ab0f72-fernet-keys\") pod \"keystone-cron-29421601-2jlb4\" (UID: \"629496e1-9961-4c1d-8f3c-9426f7ab0f72\") " pod="openstack/keystone-cron-29421601-2jlb4" Dec 09 16:01:00 crc kubenswrapper[4894]: I1209 16:01:00.166035 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rxvh2\" (UniqueName: \"kubernetes.io/projected/629496e1-9961-4c1d-8f3c-9426f7ab0f72-kube-api-access-rxvh2\") pod \"keystone-cron-29421601-2jlb4\" (UID: \"629496e1-9961-4c1d-8f3c-9426f7ab0f72\") " pod="openstack/keystone-cron-29421601-2jlb4" Dec 09 16:01:00 crc kubenswrapper[4894]: I1209 16:01:00.166064 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/629496e1-9961-4c1d-8f3c-9426f7ab0f72-config-data\") pod \"keystone-cron-29421601-2jlb4\" (UID: \"629496e1-9961-4c1d-8f3c-9426f7ab0f72\") " pod="openstack/keystone-cron-29421601-2jlb4" Dec 09 16:01:00 crc kubenswrapper[4894]: I1209 16:01:00.267442 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rxvh2\" (UniqueName: \"kubernetes.io/projected/629496e1-9961-4c1d-8f3c-9426f7ab0f72-kube-api-access-rxvh2\") pod \"keystone-cron-29421601-2jlb4\" (UID: \"629496e1-9961-4c1d-8f3c-9426f7ab0f72\") " pod="openstack/keystone-cron-29421601-2jlb4" Dec 09 16:01:00 crc kubenswrapper[4894]: I1209 16:01:00.267522 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/629496e1-9961-4c1d-8f3c-9426f7ab0f72-config-data\") pod \"keystone-cron-29421601-2jlb4\" (UID: \"629496e1-9961-4c1d-8f3c-9426f7ab0f72\") " pod="openstack/keystone-cron-29421601-2jlb4" Dec 09 16:01:00 crc kubenswrapper[4894]: I1209 16:01:00.267666 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629496e1-9961-4c1d-8f3c-9426f7ab0f72-combined-ca-bundle\") pod \"keystone-cron-29421601-2jlb4\" (UID: \"629496e1-9961-4c1d-8f3c-9426f7ab0f72\") " pod="openstack/keystone-cron-29421601-2jlb4" Dec 09 16:01:00 crc kubenswrapper[4894]: I1209 16:01:00.267734 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/629496e1-9961-4c1d-8f3c-9426f7ab0f72-fernet-keys\") pod \"keystone-cron-29421601-2jlb4\" (UID: \"629496e1-9961-4c1d-8f3c-9426f7ab0f72\") " pod="openstack/keystone-cron-29421601-2jlb4" Dec 09 16:01:00 crc kubenswrapper[4894]: I1209 16:01:00.284891 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629496e1-9961-4c1d-8f3c-9426f7ab0f72-combined-ca-bundle\") pod \"keystone-cron-29421601-2jlb4\" (UID: \"629496e1-9961-4c1d-8f3c-9426f7ab0f72\") " pod="openstack/keystone-cron-29421601-2jlb4" Dec 09 16:01:00 crc kubenswrapper[4894]: I1209 16:01:00.285270 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/629496e1-9961-4c1d-8f3c-9426f7ab0f72-fernet-keys\") pod \"keystone-cron-29421601-2jlb4\" (UID: \"629496e1-9961-4c1d-8f3c-9426f7ab0f72\") " pod="openstack/keystone-cron-29421601-2jlb4" Dec 09 16:01:00 crc kubenswrapper[4894]: I1209 16:01:00.285452 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/629496e1-9961-4c1d-8f3c-9426f7ab0f72-config-data\") pod \"keystone-cron-29421601-2jlb4\" (UID: \"629496e1-9961-4c1d-8f3c-9426f7ab0f72\") " pod="openstack/keystone-cron-29421601-2jlb4" Dec 09 16:01:00 crc kubenswrapper[4894]: I1209 16:01:00.289418 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rxvh2\" (UniqueName: \"kubernetes.io/projected/629496e1-9961-4c1d-8f3c-9426f7ab0f72-kube-api-access-rxvh2\") pod \"keystone-cron-29421601-2jlb4\" (UID: \"629496e1-9961-4c1d-8f3c-9426f7ab0f72\") " pod="openstack/keystone-cron-29421601-2jlb4" Dec 09 16:01:00 crc kubenswrapper[4894]: I1209 16:01:00.482810 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29421601-2jlb4" Dec 09 16:01:00 crc kubenswrapper[4894]: I1209 16:01:00.929998 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-cron-29421601-2jlb4"] Dec 09 16:01:00 crc kubenswrapper[4894]: W1209 16:01:00.941051 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod629496e1_9961_4c1d_8f3c_9426f7ab0f72.slice/crio-7c24e3936bbb401b0ed7f9cb3660eaac63a05fcbb2845455c139a3b6a99f8378 WatchSource:0}: Error finding container 7c24e3936bbb401b0ed7f9cb3660eaac63a05fcbb2845455c139a3b6a99f8378: Status 404 returned error can't find the container with id 7c24e3936bbb401b0ed7f9cb3660eaac63a05fcbb2845455c139a3b6a99f8378 Dec 09 16:01:01 crc kubenswrapper[4894]: I1209 16:01:01.032516 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-9b51-account-create-update-xvkxq"] Dec 09 16:01:01 crc kubenswrapper[4894]: I1209 16:01:01.040764 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-9b21-account-create-update-g5mj7"] Dec 09 16:01:01 crc kubenswrapper[4894]: I1209 16:01:01.048855 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-9b21-account-create-update-g5mj7"] Dec 09 16:01:01 crc kubenswrapper[4894]: I1209 16:01:01.055360 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-9b51-account-create-update-xvkxq"] Dec 09 16:01:01 crc kubenswrapper[4894]: I1209 16:01:01.826651 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29421601-2jlb4" event={"ID":"629496e1-9961-4c1d-8f3c-9426f7ab0f72","Type":"ContainerStarted","Data":"55266d9c3c2c73f94b6d8b302c03331b2172b1c6464a11185e38c338d73e33a3"} Dec 09 16:01:01 crc kubenswrapper[4894]: I1209 16:01:01.826974 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29421601-2jlb4" event={"ID":"629496e1-9961-4c1d-8f3c-9426f7ab0f72","Type":"ContainerStarted","Data":"7c24e3936bbb401b0ed7f9cb3660eaac63a05fcbb2845455c139a3b6a99f8378"} Dec 09 16:01:01 crc kubenswrapper[4894]: I1209 16:01:01.851384 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-cron-29421601-2jlb4" podStartSLOduration=1.851361053 podStartE2EDuration="1.851361053s" podCreationTimestamp="2025-12-09 16:01:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:01:01.846475642 +0000 UTC m=+1756.165686331" watchObservedRunningTime="2025-12-09 16:01:01.851361053 +0000 UTC m=+1756.170571722" Dec 09 16:01:02 crc kubenswrapper[4894]: I1209 16:01:02.125895 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22fdb6d9-4858-44bf-a856-04b66398260f" path="/var/lib/kubelet/pods/22fdb6d9-4858-44bf-a856-04b66398260f/volumes" Dec 09 16:01:02 crc kubenswrapper[4894]: I1209 16:01:02.126810 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="423a2183-9395-4c3f-ab1c-1816c1d31fb9" path="/var/lib/kubelet/pods/423a2183-9395-4c3f-ab1c-1816c1d31fb9/volumes" Dec 09 16:01:03 crc kubenswrapper[4894]: I1209 16:01:03.843551 4894 generic.go:334] "Generic (PLEG): container finished" podID="629496e1-9961-4c1d-8f3c-9426f7ab0f72" containerID="55266d9c3c2c73f94b6d8b302c03331b2172b1c6464a11185e38c338d73e33a3" exitCode=0 Dec 09 16:01:03 crc kubenswrapper[4894]: I1209 16:01:03.843864 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29421601-2jlb4" event={"ID":"629496e1-9961-4c1d-8f3c-9426f7ab0f72","Type":"ContainerDied","Data":"55266d9c3c2c73f94b6d8b302c03331b2172b1c6464a11185e38c338d73e33a3"} Dec 09 16:01:05 crc kubenswrapper[4894]: I1209 16:01:05.183175 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29421601-2jlb4" Dec 09 16:01:05 crc kubenswrapper[4894]: I1209 16:01:05.367945 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/629496e1-9961-4c1d-8f3c-9426f7ab0f72-fernet-keys\") pod \"629496e1-9961-4c1d-8f3c-9426f7ab0f72\" (UID: \"629496e1-9961-4c1d-8f3c-9426f7ab0f72\") " Dec 09 16:01:05 crc kubenswrapper[4894]: I1209 16:01:05.368184 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629496e1-9961-4c1d-8f3c-9426f7ab0f72-combined-ca-bundle\") pod \"629496e1-9961-4c1d-8f3c-9426f7ab0f72\" (UID: \"629496e1-9961-4c1d-8f3c-9426f7ab0f72\") " Dec 09 16:01:05 crc kubenswrapper[4894]: I1209 16:01:05.368292 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rxvh2\" (UniqueName: \"kubernetes.io/projected/629496e1-9961-4c1d-8f3c-9426f7ab0f72-kube-api-access-rxvh2\") pod \"629496e1-9961-4c1d-8f3c-9426f7ab0f72\" (UID: \"629496e1-9961-4c1d-8f3c-9426f7ab0f72\") " Dec 09 16:01:05 crc kubenswrapper[4894]: I1209 16:01:05.368374 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/629496e1-9961-4c1d-8f3c-9426f7ab0f72-config-data\") pod \"629496e1-9961-4c1d-8f3c-9426f7ab0f72\" (UID: \"629496e1-9961-4c1d-8f3c-9426f7ab0f72\") " Dec 09 16:01:05 crc kubenswrapper[4894]: I1209 16:01:05.374238 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/629496e1-9961-4c1d-8f3c-9426f7ab0f72-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "629496e1-9961-4c1d-8f3c-9426f7ab0f72" (UID: "629496e1-9961-4c1d-8f3c-9426f7ab0f72"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:01:05 crc kubenswrapper[4894]: I1209 16:01:05.379392 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/629496e1-9961-4c1d-8f3c-9426f7ab0f72-kube-api-access-rxvh2" (OuterVolumeSpecName: "kube-api-access-rxvh2") pod "629496e1-9961-4c1d-8f3c-9426f7ab0f72" (UID: "629496e1-9961-4c1d-8f3c-9426f7ab0f72"). InnerVolumeSpecName "kube-api-access-rxvh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:01:05 crc kubenswrapper[4894]: I1209 16:01:05.395788 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/629496e1-9961-4c1d-8f3c-9426f7ab0f72-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "629496e1-9961-4c1d-8f3c-9426f7ab0f72" (UID: "629496e1-9961-4c1d-8f3c-9426f7ab0f72"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:01:05 crc kubenswrapper[4894]: I1209 16:01:05.416492 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/629496e1-9961-4c1d-8f3c-9426f7ab0f72-config-data" (OuterVolumeSpecName: "config-data") pod "629496e1-9961-4c1d-8f3c-9426f7ab0f72" (UID: "629496e1-9961-4c1d-8f3c-9426f7ab0f72"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:01:05 crc kubenswrapper[4894]: I1209 16:01:05.470078 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/629496e1-9961-4c1d-8f3c-9426f7ab0f72-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:01:05 crc kubenswrapper[4894]: I1209 16:01:05.470290 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rxvh2\" (UniqueName: \"kubernetes.io/projected/629496e1-9961-4c1d-8f3c-9426f7ab0f72-kube-api-access-rxvh2\") on node \"crc\" DevicePath \"\"" Dec 09 16:01:05 crc kubenswrapper[4894]: I1209 16:01:05.470382 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/629496e1-9961-4c1d-8f3c-9426f7ab0f72-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 16:01:05 crc kubenswrapper[4894]: I1209 16:01:05.470456 4894 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/629496e1-9961-4c1d-8f3c-9426f7ab0f72-fernet-keys\") on node \"crc\" DevicePath \"\"" Dec 09 16:01:05 crc kubenswrapper[4894]: I1209 16:01:05.867217 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-cron-29421601-2jlb4" event={"ID":"629496e1-9961-4c1d-8f3c-9426f7ab0f72","Type":"ContainerDied","Data":"7c24e3936bbb401b0ed7f9cb3660eaac63a05fcbb2845455c139a3b6a99f8378"} Dec 09 16:01:05 crc kubenswrapper[4894]: I1209 16:01:05.867462 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-cron-29421601-2jlb4" Dec 09 16:01:05 crc kubenswrapper[4894]: I1209 16:01:05.867479 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7c24e3936bbb401b0ed7f9cb3660eaac63a05fcbb2845455c139a3b6a99f8378" Dec 09 16:01:08 crc kubenswrapper[4894]: I1209 16:01:08.897815 4894 generic.go:334] "Generic (PLEG): container finished" podID="94a3b079-5c34-4ef3-9262-eb741ef7910b" containerID="ad356faf9f26c46787e5421cec638ad9dee018928d9909ef8762fe781d29ed10" exitCode=0 Dec 09 16:01:08 crc kubenswrapper[4894]: I1209 16:01:08.897885 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp" event={"ID":"94a3b079-5c34-4ef3-9262-eb741ef7910b","Type":"ContainerDied","Data":"ad356faf9f26c46787e5421cec638ad9dee018928d9909ef8762fe781d29ed10"} Dec 09 16:01:10 crc kubenswrapper[4894]: I1209 16:01:10.365430 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp" Dec 09 16:01:10 crc kubenswrapper[4894]: I1209 16:01:10.475656 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94a3b079-5c34-4ef3-9262-eb741ef7910b-ssh-key\") pod \"94a3b079-5c34-4ef3-9262-eb741ef7910b\" (UID: \"94a3b079-5c34-4ef3-9262-eb741ef7910b\") " Dec 09 16:01:10 crc kubenswrapper[4894]: I1209 16:01:10.475716 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94a3b079-5c34-4ef3-9262-eb741ef7910b-inventory\") pod \"94a3b079-5c34-4ef3-9262-eb741ef7910b\" (UID: \"94a3b079-5c34-4ef3-9262-eb741ef7910b\") " Dec 09 16:01:10 crc kubenswrapper[4894]: I1209 16:01:10.475793 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-84znz\" (UniqueName: \"kubernetes.io/projected/94a3b079-5c34-4ef3-9262-eb741ef7910b-kube-api-access-84znz\") pod \"94a3b079-5c34-4ef3-9262-eb741ef7910b\" (UID: \"94a3b079-5c34-4ef3-9262-eb741ef7910b\") " Dec 09 16:01:10 crc kubenswrapper[4894]: I1209 16:01:10.481863 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/94a3b079-5c34-4ef3-9262-eb741ef7910b-kube-api-access-84znz" (OuterVolumeSpecName: "kube-api-access-84znz") pod "94a3b079-5c34-4ef3-9262-eb741ef7910b" (UID: "94a3b079-5c34-4ef3-9262-eb741ef7910b"). InnerVolumeSpecName "kube-api-access-84znz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:01:10 crc kubenswrapper[4894]: I1209 16:01:10.508684 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94a3b079-5c34-4ef3-9262-eb741ef7910b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "94a3b079-5c34-4ef3-9262-eb741ef7910b" (UID: "94a3b079-5c34-4ef3-9262-eb741ef7910b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:01:10 crc kubenswrapper[4894]: I1209 16:01:10.512102 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/94a3b079-5c34-4ef3-9262-eb741ef7910b-inventory" (OuterVolumeSpecName: "inventory") pod "94a3b079-5c34-4ef3-9262-eb741ef7910b" (UID: "94a3b079-5c34-4ef3-9262-eb741ef7910b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:01:10 crc kubenswrapper[4894]: I1209 16:01:10.578105 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/94a3b079-5c34-4ef3-9262-eb741ef7910b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:01:10 crc kubenswrapper[4894]: I1209 16:01:10.578143 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/94a3b079-5c34-4ef3-9262-eb741ef7910b-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 16:01:10 crc kubenswrapper[4894]: I1209 16:01:10.578153 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-84znz\" (UniqueName: \"kubernetes.io/projected/94a3b079-5c34-4ef3-9262-eb741ef7910b-kube-api-access-84znz\") on node \"crc\" DevicePath \"\"" Dec 09 16:01:10 crc kubenswrapper[4894]: I1209 16:01:10.921104 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp" event={"ID":"94a3b079-5c34-4ef3-9262-eb741ef7910b","Type":"ContainerDied","Data":"188fc594cea51c37032d8464419192d1057841d3721f2c2500b8b57d12fb1d54"} Dec 09 16:01:10 crc kubenswrapper[4894]: I1209 16:01:10.921151 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp" Dec 09 16:01:10 crc kubenswrapper[4894]: I1209 16:01:10.921163 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="188fc594cea51c37032d8464419192d1057841d3721f2c2500b8b57d12fb1d54" Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.003780 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-tzfrh"] Dec 09 16:01:11 crc kubenswrapper[4894]: E1209 16:01:11.004139 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="94a3b079-5c34-4ef3-9262-eb741ef7910b" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.004154 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="94a3b079-5c34-4ef3-9262-eb741ef7910b" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:01:11 crc kubenswrapper[4894]: E1209 16:01:11.004180 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="629496e1-9961-4c1d-8f3c-9426f7ab0f72" containerName="keystone-cron" Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.004187 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="629496e1-9961-4c1d-8f3c-9426f7ab0f72" containerName="keystone-cron" Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.004348 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="629496e1-9961-4c1d-8f3c-9426f7ab0f72" containerName="keystone-cron" Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.004373 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="94a3b079-5c34-4ef3-9262-eb741ef7910b" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.005074 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-tzfrh" Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.012619 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.013839 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.015272 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.015349 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.033135 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-tzfrh"] Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.089067 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cfkxj\" (UniqueName: \"kubernetes.io/projected/7dec777f-52ed-4e72-8510-352d8da796f7-kube-api-access-cfkxj\") pod \"ssh-known-hosts-edpm-deployment-tzfrh\" (UID: \"7dec777f-52ed-4e72-8510-352d8da796f7\") " pod="openstack/ssh-known-hosts-edpm-deployment-tzfrh" Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.089262 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7dec777f-52ed-4e72-8510-352d8da796f7-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-tzfrh\" (UID: \"7dec777f-52ed-4e72-8510-352d8da796f7\") " pod="openstack/ssh-known-hosts-edpm-deployment-tzfrh" Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.089346 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7dec777f-52ed-4e72-8510-352d8da796f7-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-tzfrh\" (UID: \"7dec777f-52ed-4e72-8510-352d8da796f7\") " pod="openstack/ssh-known-hosts-edpm-deployment-tzfrh" Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.106908 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 16:01:11 crc kubenswrapper[4894]: E1209 16:01:11.107295 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.191274 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cfkxj\" (UniqueName: \"kubernetes.io/projected/7dec777f-52ed-4e72-8510-352d8da796f7-kube-api-access-cfkxj\") pod \"ssh-known-hosts-edpm-deployment-tzfrh\" (UID: \"7dec777f-52ed-4e72-8510-352d8da796f7\") " pod="openstack/ssh-known-hosts-edpm-deployment-tzfrh" Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.191446 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7dec777f-52ed-4e72-8510-352d8da796f7-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-tzfrh\" (UID: \"7dec777f-52ed-4e72-8510-352d8da796f7\") " pod="openstack/ssh-known-hosts-edpm-deployment-tzfrh" Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.191521 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7dec777f-52ed-4e72-8510-352d8da796f7-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-tzfrh\" (UID: \"7dec777f-52ed-4e72-8510-352d8da796f7\") " pod="openstack/ssh-known-hosts-edpm-deployment-tzfrh" Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.196626 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7dec777f-52ed-4e72-8510-352d8da796f7-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-tzfrh\" (UID: \"7dec777f-52ed-4e72-8510-352d8da796f7\") " pod="openstack/ssh-known-hosts-edpm-deployment-tzfrh" Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.199470 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7dec777f-52ed-4e72-8510-352d8da796f7-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-tzfrh\" (UID: \"7dec777f-52ed-4e72-8510-352d8da796f7\") " pod="openstack/ssh-known-hosts-edpm-deployment-tzfrh" Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.215228 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cfkxj\" (UniqueName: \"kubernetes.io/projected/7dec777f-52ed-4e72-8510-352d8da796f7-kube-api-access-cfkxj\") pod \"ssh-known-hosts-edpm-deployment-tzfrh\" (UID: \"7dec777f-52ed-4e72-8510-352d8da796f7\") " pod="openstack/ssh-known-hosts-edpm-deployment-tzfrh" Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.338884 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-tzfrh" Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.873892 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-tzfrh"] Dec 09 16:01:11 crc kubenswrapper[4894]: I1209 16:01:11.932436 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-tzfrh" event={"ID":"7dec777f-52ed-4e72-8510-352d8da796f7","Type":"ContainerStarted","Data":"ab9a455a239c5b2f390361836daf4cde40ba4dc8ab5fb1d81f34ae0cc61e0cb9"} Dec 09 16:01:13 crc kubenswrapper[4894]: I1209 16:01:13.957821 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-tzfrh" event={"ID":"7dec777f-52ed-4e72-8510-352d8da796f7","Type":"ContainerStarted","Data":"18e7396abc35ba282dc17c5c9603c81460f6a84975fcfede4834c7f155539f72"} Dec 09 16:01:13 crc kubenswrapper[4894]: I1209 16:01:13.980255 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-tzfrh" podStartSLOduration=3.177780184 podStartE2EDuration="3.980238979s" podCreationTimestamp="2025-12-09 16:01:10 +0000 UTC" firstStartedPulling="2025-12-09 16:01:11.870110857 +0000 UTC m=+1766.189321526" lastFinishedPulling="2025-12-09 16:01:12.672569642 +0000 UTC m=+1766.991780321" observedRunningTime="2025-12-09 16:01:13.977335131 +0000 UTC m=+1768.296545810" watchObservedRunningTime="2025-12-09 16:01:13.980238979 +0000 UTC m=+1768.299449648" Dec 09 16:01:20 crc kubenswrapper[4894]: I1209 16:01:20.008183 4894 generic.go:334] "Generic (PLEG): container finished" podID="7dec777f-52ed-4e72-8510-352d8da796f7" containerID="18e7396abc35ba282dc17c5c9603c81460f6a84975fcfede4834c7f155539f72" exitCode=0 Dec 09 16:01:20 crc kubenswrapper[4894]: I1209 16:01:20.008246 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-tzfrh" event={"ID":"7dec777f-52ed-4e72-8510-352d8da796f7","Type":"ContainerDied","Data":"18e7396abc35ba282dc17c5c9603c81460f6a84975fcfede4834c7f155539f72"} Dec 09 16:01:21 crc kubenswrapper[4894]: I1209 16:01:21.466628 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-tzfrh" Dec 09 16:01:21 crc kubenswrapper[4894]: I1209 16:01:21.481079 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7dec777f-52ed-4e72-8510-352d8da796f7-inventory-0\") pod \"7dec777f-52ed-4e72-8510-352d8da796f7\" (UID: \"7dec777f-52ed-4e72-8510-352d8da796f7\") " Dec 09 16:01:21 crc kubenswrapper[4894]: I1209 16:01:21.481189 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7dec777f-52ed-4e72-8510-352d8da796f7-ssh-key-openstack-edpm-ipam\") pod \"7dec777f-52ed-4e72-8510-352d8da796f7\" (UID: \"7dec777f-52ed-4e72-8510-352d8da796f7\") " Dec 09 16:01:21 crc kubenswrapper[4894]: I1209 16:01:21.481468 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfkxj\" (UniqueName: \"kubernetes.io/projected/7dec777f-52ed-4e72-8510-352d8da796f7-kube-api-access-cfkxj\") pod \"7dec777f-52ed-4e72-8510-352d8da796f7\" (UID: \"7dec777f-52ed-4e72-8510-352d8da796f7\") " Dec 09 16:01:21 crc kubenswrapper[4894]: I1209 16:01:21.491150 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7dec777f-52ed-4e72-8510-352d8da796f7-kube-api-access-cfkxj" (OuterVolumeSpecName: "kube-api-access-cfkxj") pod "7dec777f-52ed-4e72-8510-352d8da796f7" (UID: "7dec777f-52ed-4e72-8510-352d8da796f7"). InnerVolumeSpecName "kube-api-access-cfkxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:01:21 crc kubenswrapper[4894]: I1209 16:01:21.513709 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dec777f-52ed-4e72-8510-352d8da796f7-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "7dec777f-52ed-4e72-8510-352d8da796f7" (UID: "7dec777f-52ed-4e72-8510-352d8da796f7"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:01:21 crc kubenswrapper[4894]: I1209 16:01:21.525311 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7dec777f-52ed-4e72-8510-352d8da796f7-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "7dec777f-52ed-4e72-8510-352d8da796f7" (UID: "7dec777f-52ed-4e72-8510-352d8da796f7"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:01:21 crc kubenswrapper[4894]: I1209 16:01:21.583499 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfkxj\" (UniqueName: \"kubernetes.io/projected/7dec777f-52ed-4e72-8510-352d8da796f7-kube-api-access-cfkxj\") on node \"crc\" DevicePath \"\"" Dec 09 16:01:21 crc kubenswrapper[4894]: I1209 16:01:21.583537 4894 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/7dec777f-52ed-4e72-8510-352d8da796f7-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 09 16:01:21 crc kubenswrapper[4894]: I1209 16:01:21.583547 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/7dec777f-52ed-4e72-8510-352d8da796f7-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.027004 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-tzfrh" event={"ID":"7dec777f-52ed-4e72-8510-352d8da796f7","Type":"ContainerDied","Data":"ab9a455a239c5b2f390361836daf4cde40ba4dc8ab5fb1d81f34ae0cc61e0cb9"} Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.027053 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ab9a455a239c5b2f390361836daf4cde40ba4dc8ab5fb1d81f34ae0cc61e0cb9" Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.027122 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-tzfrh" Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.116313 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4"] Dec 09 16:01:22 crc kubenswrapper[4894]: E1209 16:01:22.116715 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7dec777f-52ed-4e72-8510-352d8da796f7" containerName="ssh-known-hosts-edpm-deployment" Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.116732 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7dec777f-52ed-4e72-8510-352d8da796f7" containerName="ssh-known-hosts-edpm-deployment" Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.116951 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7dec777f-52ed-4e72-8510-352d8da796f7" containerName="ssh-known-hosts-edpm-deployment" Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.117686 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4" Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.120773 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4"] Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.127169 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.127447 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.128023 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.128081 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.192491 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gvfz\" (UniqueName: \"kubernetes.io/projected/22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1-kube-api-access-5gvfz\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ntpv4\" (UID: \"22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4" Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.192656 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ntpv4\" (UID: \"22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4" Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.192958 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ntpv4\" (UID: \"22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4" Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.294781 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ntpv4\" (UID: \"22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4" Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.294885 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ntpv4\" (UID: \"22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4" Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.294960 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gvfz\" (UniqueName: \"kubernetes.io/projected/22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1-kube-api-access-5gvfz\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ntpv4\" (UID: \"22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4" Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.298948 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ntpv4\" (UID: \"22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4" Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.300485 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ntpv4\" (UID: \"22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4" Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.313814 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gvfz\" (UniqueName: \"kubernetes.io/projected/22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1-kube-api-access-5gvfz\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-ntpv4\" (UID: \"22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4" Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.434995 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4" Dec 09 16:01:22 crc kubenswrapper[4894]: I1209 16:01:22.991227 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4"] Dec 09 16:01:23 crc kubenswrapper[4894]: I1209 16:01:23.034380 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4" event={"ID":"22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1","Type":"ContainerStarted","Data":"1286d4b0853750bec43044804f1ba976992fbc4b6e942630180977e7ae13b656"} Dec 09 16:01:24 crc kubenswrapper[4894]: I1209 16:01:24.044703 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4" event={"ID":"22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1","Type":"ContainerStarted","Data":"7cbc03646abf25ca6598e20f165df3e84d297fda8fcce968ff2ba15bda949c9a"} Dec 09 16:01:24 crc kubenswrapper[4894]: I1209 16:01:24.066761 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4" podStartSLOduration=1.545961785 podStartE2EDuration="2.066738976s" podCreationTimestamp="2025-12-09 16:01:22 +0000 UTC" firstStartedPulling="2025-12-09 16:01:23.000927461 +0000 UTC m=+1777.320138130" lastFinishedPulling="2025-12-09 16:01:23.521704652 +0000 UTC m=+1777.840915321" observedRunningTime="2025-12-09 16:01:24.058580226 +0000 UTC m=+1778.377790895" watchObservedRunningTime="2025-12-09 16:01:24.066738976 +0000 UTC m=+1778.385949645" Dec 09 16:01:24 crc kubenswrapper[4894]: I1209 16:01:24.106749 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 16:01:24 crc kubenswrapper[4894]: E1209 16:01:24.107222 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:01:33 crc kubenswrapper[4894]: I1209 16:01:33.142515 4894 generic.go:334] "Generic (PLEG): container finished" podID="22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1" containerID="7cbc03646abf25ca6598e20f165df3e84d297fda8fcce968ff2ba15bda949c9a" exitCode=0 Dec 09 16:01:33 crc kubenswrapper[4894]: I1209 16:01:33.142922 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4" event={"ID":"22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1","Type":"ContainerDied","Data":"7cbc03646abf25ca6598e20f165df3e84d297fda8fcce968ff2ba15bda949c9a"} Dec 09 16:01:34 crc kubenswrapper[4894]: I1209 16:01:34.546078 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4" Dec 09 16:01:34 crc kubenswrapper[4894]: I1209 16:01:34.740345 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gvfz\" (UniqueName: \"kubernetes.io/projected/22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1-kube-api-access-5gvfz\") pod \"22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1\" (UID: \"22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1\") " Dec 09 16:01:34 crc kubenswrapper[4894]: I1209 16:01:34.740422 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1-ssh-key\") pod \"22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1\" (UID: \"22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1\") " Dec 09 16:01:34 crc kubenswrapper[4894]: I1209 16:01:34.740652 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1-inventory\") pod \"22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1\" (UID: \"22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1\") " Dec 09 16:01:34 crc kubenswrapper[4894]: I1209 16:01:34.746993 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1-kube-api-access-5gvfz" (OuterVolumeSpecName: "kube-api-access-5gvfz") pod "22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1" (UID: "22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1"). InnerVolumeSpecName "kube-api-access-5gvfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:01:34 crc kubenswrapper[4894]: I1209 16:01:34.771982 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1-inventory" (OuterVolumeSpecName: "inventory") pod "22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1" (UID: "22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:01:34 crc kubenswrapper[4894]: I1209 16:01:34.775866 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1" (UID: "22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:01:34 crc kubenswrapper[4894]: I1209 16:01:34.843034 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 16:01:34 crc kubenswrapper[4894]: I1209 16:01:34.843085 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gvfz\" (UniqueName: \"kubernetes.io/projected/22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1-kube-api-access-5gvfz\") on node \"crc\" DevicePath \"\"" Dec 09 16:01:34 crc kubenswrapper[4894]: I1209 16:01:34.843105 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.107387 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 16:01:35 crc kubenswrapper[4894]: E1209 16:01:35.107963 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.161328 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4" event={"ID":"22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1","Type":"ContainerDied","Data":"1286d4b0853750bec43044804f1ba976992fbc4b6e942630180977e7ae13b656"} Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.161376 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1286d4b0853750bec43044804f1ba976992fbc4b6e942630180977e7ae13b656" Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.161436 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4" Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.264807 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n"] Dec 09 16:01:35 crc kubenswrapper[4894]: E1209 16:01:35.265318 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.265346 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.265561 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.266312 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n" Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.270077 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.272452 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.272799 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.273139 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.279853 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n"] Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.454466 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n\" (UID: \"829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n" Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.455245 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n\" (UID: \"829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n" Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.455348 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwgx2\" (UniqueName: \"kubernetes.io/projected/829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2-kube-api-access-mwgx2\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n\" (UID: \"829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n" Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.558077 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n\" (UID: \"829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n" Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.558187 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwgx2\" (UniqueName: \"kubernetes.io/projected/829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2-kube-api-access-mwgx2\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n\" (UID: \"829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n" Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.558315 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n\" (UID: \"829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n" Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.565285 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n\" (UID: \"829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n" Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.565338 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n\" (UID: \"829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n" Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.581166 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwgx2\" (UniqueName: \"kubernetes.io/projected/829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2-kube-api-access-mwgx2\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n\" (UID: \"829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n" Dec 09 16:01:35 crc kubenswrapper[4894]: I1209 16:01:35.590423 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n" Dec 09 16:01:36 crc kubenswrapper[4894]: I1209 16:01:36.077831 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-b56ck"] Dec 09 16:01:36 crc kubenswrapper[4894]: I1209 16:01:36.128239 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-b56ck"] Dec 09 16:01:36 crc kubenswrapper[4894]: I1209 16:01:36.159708 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n"] Dec 09 16:01:36 crc kubenswrapper[4894]: I1209 16:01:36.175887 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n" event={"ID":"829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2","Type":"ContainerStarted","Data":"1832d0b93459112dc6a9a4bcc9b2c2b165f69e5587a0153e3f8097364538d6a1"} Dec 09 16:01:38 crc kubenswrapper[4894]: I1209 16:01:38.117004 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8ed12da4-c785-4d59-a3d7-c485b97013bc" path="/var/lib/kubelet/pods/8ed12da4-c785-4d59-a3d7-c485b97013bc/volumes" Dec 09 16:01:38 crc kubenswrapper[4894]: I1209 16:01:38.199335 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n" event={"ID":"829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2","Type":"ContainerStarted","Data":"932d21281cf6b4c3932dafbd6482a4f38deb9a110fe81c7256e28761b3dc7591"} Dec 09 16:01:38 crc kubenswrapper[4894]: I1209 16:01:38.225709 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n" podStartSLOduration=1.847694677 podStartE2EDuration="3.225687088s" podCreationTimestamp="2025-12-09 16:01:35 +0000 UTC" firstStartedPulling="2025-12-09 16:01:36.169224671 +0000 UTC m=+1790.488435340" lastFinishedPulling="2025-12-09 16:01:37.547217072 +0000 UTC m=+1791.866427751" observedRunningTime="2025-12-09 16:01:38.220254822 +0000 UTC m=+1792.539465491" watchObservedRunningTime="2025-12-09 16:01:38.225687088 +0000 UTC m=+1792.544897767" Dec 09 16:01:46 crc kubenswrapper[4894]: I1209 16:01:46.116999 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 16:01:46 crc kubenswrapper[4894]: E1209 16:01:46.117860 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:01:48 crc kubenswrapper[4894]: I1209 16:01:48.300099 4894 generic.go:334] "Generic (PLEG): container finished" podID="829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2" containerID="932d21281cf6b4c3932dafbd6482a4f38deb9a110fe81c7256e28761b3dc7591" exitCode=0 Dec 09 16:01:48 crc kubenswrapper[4894]: I1209 16:01:48.300180 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n" event={"ID":"829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2","Type":"ContainerDied","Data":"932d21281cf6b4c3932dafbd6482a4f38deb9a110fe81c7256e28761b3dc7591"} Dec 09 16:01:49 crc kubenswrapper[4894]: I1209 16:01:49.723963 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n" Dec 09 16:01:49 crc kubenswrapper[4894]: I1209 16:01:49.835828 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2-ssh-key\") pod \"829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2\" (UID: \"829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2\") " Dec 09 16:01:49 crc kubenswrapper[4894]: I1209 16:01:49.836227 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwgx2\" (UniqueName: \"kubernetes.io/projected/829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2-kube-api-access-mwgx2\") pod \"829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2\" (UID: \"829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2\") " Dec 09 16:01:49 crc kubenswrapper[4894]: I1209 16:01:49.836359 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2-inventory\") pod \"829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2\" (UID: \"829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2\") " Dec 09 16:01:49 crc kubenswrapper[4894]: I1209 16:01:49.841568 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2-kube-api-access-mwgx2" (OuterVolumeSpecName: "kube-api-access-mwgx2") pod "829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2" (UID: "829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2"). InnerVolumeSpecName "kube-api-access-mwgx2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:01:49 crc kubenswrapper[4894]: I1209 16:01:49.863972 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2" (UID: "829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:01:49 crc kubenswrapper[4894]: I1209 16:01:49.886308 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2-inventory" (OuterVolumeSpecName: "inventory") pod "829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2" (UID: "829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:01:49 crc kubenswrapper[4894]: I1209 16:01:49.939054 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 16:01:49 crc kubenswrapper[4894]: I1209 16:01:49.939101 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:01:49 crc kubenswrapper[4894]: I1209 16:01:49.939117 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwgx2\" (UniqueName: \"kubernetes.io/projected/829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2-kube-api-access-mwgx2\") on node \"crc\" DevicePath \"\"" Dec 09 16:01:50 crc kubenswrapper[4894]: I1209 16:01:50.317877 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n" event={"ID":"829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2","Type":"ContainerDied","Data":"1832d0b93459112dc6a9a4bcc9b2c2b165f69e5587a0153e3f8097364538d6a1"} Dec 09 16:01:50 crc kubenswrapper[4894]: I1209 16:01:50.317913 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1832d0b93459112dc6a9a4bcc9b2c2b165f69e5587a0153e3f8097364538d6a1" Dec 09 16:01:50 crc kubenswrapper[4894]: I1209 16:01:50.317923 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n" Dec 09 16:01:56 crc kubenswrapper[4894]: I1209 16:01:56.042401 4894 scope.go:117] "RemoveContainer" containerID="a622ced4830d636677017d89b88e3a2170266e342db1cbe48a5537d668c423cd" Dec 09 16:01:56 crc kubenswrapper[4894]: I1209 16:01:56.070680 4894 scope.go:117] "RemoveContainer" containerID="ebb2d23f3bfa50765824c94ae14d468a88a6a0b774c71abcf01a4eb038186069" Dec 09 16:01:56 crc kubenswrapper[4894]: I1209 16:01:56.112227 4894 scope.go:117] "RemoveContainer" containerID="eb0e84f294bc81396d21a68308cbb6daf5f04026d580ce87f6bd2297b9c4fb4e" Dec 09 16:01:59 crc kubenswrapper[4894]: I1209 16:01:59.076631 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-w26bv"] Dec 09 16:01:59 crc kubenswrapper[4894]: I1209 16:01:59.083658 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-w26bv"] Dec 09 16:02:00 crc kubenswrapper[4894]: I1209 16:02:00.117869 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e2f0b872-8eba-4309-94c8-4dfe893cfc3c" path="/var/lib/kubelet/pods/e2f0b872-8eba-4309-94c8-4dfe893cfc3c/volumes" Dec 09 16:02:01 crc kubenswrapper[4894]: I1209 16:02:01.038856 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-scszh"] Dec 09 16:02:01 crc kubenswrapper[4894]: I1209 16:02:01.059975 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-scszh"] Dec 09 16:02:01 crc kubenswrapper[4894]: I1209 16:02:01.106713 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 16:02:01 crc kubenswrapper[4894]: E1209 16:02:01.109031 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:02:02 crc kubenswrapper[4894]: I1209 16:02:02.116787 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0f261fad-179c-4f25-bc42-44691e48be50" path="/var/lib/kubelet/pods/0f261fad-179c-4f25-bc42-44691e48be50/volumes" Dec 09 16:02:12 crc kubenswrapper[4894]: I1209 16:02:12.106761 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 16:02:12 crc kubenswrapper[4894]: E1209 16:02:12.107681 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:02:26 crc kubenswrapper[4894]: I1209 16:02:26.118984 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 16:02:26 crc kubenswrapper[4894]: I1209 16:02:26.682794 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerStarted","Data":"b62d7b69e0742d175c7cec7191590d1fa9926a45adb97c2aa898cb0cc3a77b56"} Dec 09 16:02:44 crc kubenswrapper[4894]: I1209 16:02:44.050118 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-kwtnn"] Dec 09 16:02:44 crc kubenswrapper[4894]: I1209 16:02:44.056873 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-kwtnn"] Dec 09 16:02:44 crc kubenswrapper[4894]: I1209 16:02:44.119433 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95bff0ec-6b68-439a-b478-de6549452662" path="/var/lib/kubelet/pods/95bff0ec-6b68-439a-b478-de6549452662/volumes" Dec 09 16:02:56 crc kubenswrapper[4894]: I1209 16:02:56.235257 4894 scope.go:117] "RemoveContainer" containerID="05878ba3c9a9a398a2afffd857ceedb589bc02e084d330c155e3c537a8230aa9" Dec 09 16:02:56 crc kubenswrapper[4894]: I1209 16:02:56.280870 4894 scope.go:117] "RemoveContainer" containerID="389f6a7420333dcd236202ca16cd594b4e9bb3dd3e8434999e9784a6d984c7c9" Dec 09 16:02:56 crc kubenswrapper[4894]: I1209 16:02:56.352232 4894 scope.go:117] "RemoveContainer" containerID="29e9cc197d569126cf8be32f02028738bb59d1c4167b6af1d9d999aecd711a2e" Dec 09 16:03:13 crc kubenswrapper[4894]: I1209 16:03:13.797345 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7rmdd"] Dec 09 16:03:13 crc kubenswrapper[4894]: E1209 16:03:13.805272 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:03:13 crc kubenswrapper[4894]: I1209 16:03:13.805321 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:03:13 crc kubenswrapper[4894]: I1209 16:03:13.805717 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:03:13 crc kubenswrapper[4894]: I1209 16:03:13.808083 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7rmdd" Dec 09 16:03:13 crc kubenswrapper[4894]: I1209 16:03:13.814880 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7rmdd"] Dec 09 16:03:13 crc kubenswrapper[4894]: I1209 16:03:13.951783 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59467bee-13bb-4c8c-8617-b485ac0d0de0-catalog-content\") pod \"certified-operators-7rmdd\" (UID: \"59467bee-13bb-4c8c-8617-b485ac0d0de0\") " pod="openshift-marketplace/certified-operators-7rmdd" Dec 09 16:03:13 crc kubenswrapper[4894]: I1209 16:03:13.951926 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z2kw7\" (UniqueName: \"kubernetes.io/projected/59467bee-13bb-4c8c-8617-b485ac0d0de0-kube-api-access-z2kw7\") pod \"certified-operators-7rmdd\" (UID: \"59467bee-13bb-4c8c-8617-b485ac0d0de0\") " pod="openshift-marketplace/certified-operators-7rmdd" Dec 09 16:03:13 crc kubenswrapper[4894]: I1209 16:03:13.952018 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59467bee-13bb-4c8c-8617-b485ac0d0de0-utilities\") pod \"certified-operators-7rmdd\" (UID: \"59467bee-13bb-4c8c-8617-b485ac0d0de0\") " pod="openshift-marketplace/certified-operators-7rmdd" Dec 09 16:03:14 crc kubenswrapper[4894]: I1209 16:03:14.053959 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59467bee-13bb-4c8c-8617-b485ac0d0de0-catalog-content\") pod \"certified-operators-7rmdd\" (UID: \"59467bee-13bb-4c8c-8617-b485ac0d0de0\") " pod="openshift-marketplace/certified-operators-7rmdd" Dec 09 16:03:14 crc kubenswrapper[4894]: I1209 16:03:14.054048 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z2kw7\" (UniqueName: \"kubernetes.io/projected/59467bee-13bb-4c8c-8617-b485ac0d0de0-kube-api-access-z2kw7\") pod \"certified-operators-7rmdd\" (UID: \"59467bee-13bb-4c8c-8617-b485ac0d0de0\") " pod="openshift-marketplace/certified-operators-7rmdd" Dec 09 16:03:14 crc kubenswrapper[4894]: I1209 16:03:14.054126 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59467bee-13bb-4c8c-8617-b485ac0d0de0-utilities\") pod \"certified-operators-7rmdd\" (UID: \"59467bee-13bb-4c8c-8617-b485ac0d0de0\") " pod="openshift-marketplace/certified-operators-7rmdd" Dec 09 16:03:14 crc kubenswrapper[4894]: I1209 16:03:14.054614 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59467bee-13bb-4c8c-8617-b485ac0d0de0-utilities\") pod \"certified-operators-7rmdd\" (UID: \"59467bee-13bb-4c8c-8617-b485ac0d0de0\") " pod="openshift-marketplace/certified-operators-7rmdd" Dec 09 16:03:14 crc kubenswrapper[4894]: I1209 16:03:14.054677 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59467bee-13bb-4c8c-8617-b485ac0d0de0-catalog-content\") pod \"certified-operators-7rmdd\" (UID: \"59467bee-13bb-4c8c-8617-b485ac0d0de0\") " pod="openshift-marketplace/certified-operators-7rmdd" Dec 09 16:03:14 crc kubenswrapper[4894]: I1209 16:03:14.078057 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z2kw7\" (UniqueName: \"kubernetes.io/projected/59467bee-13bb-4c8c-8617-b485ac0d0de0-kube-api-access-z2kw7\") pod \"certified-operators-7rmdd\" (UID: \"59467bee-13bb-4c8c-8617-b485ac0d0de0\") " pod="openshift-marketplace/certified-operators-7rmdd" Dec 09 16:03:14 crc kubenswrapper[4894]: I1209 16:03:14.146288 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7rmdd" Dec 09 16:03:14 crc kubenswrapper[4894]: I1209 16:03:14.708610 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7rmdd"] Dec 09 16:03:15 crc kubenswrapper[4894]: I1209 16:03:15.160199 4894 generic.go:334] "Generic (PLEG): container finished" podID="59467bee-13bb-4c8c-8617-b485ac0d0de0" containerID="d4c83d52bd72e094875513bbc2f3405469de0ca7f7ea19d34f7e5882f5afcd8d" exitCode=0 Dec 09 16:03:15 crc kubenswrapper[4894]: I1209 16:03:15.160281 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7rmdd" event={"ID":"59467bee-13bb-4c8c-8617-b485ac0d0de0","Type":"ContainerDied","Data":"d4c83d52bd72e094875513bbc2f3405469de0ca7f7ea19d34f7e5882f5afcd8d"} Dec 09 16:03:15 crc kubenswrapper[4894]: I1209 16:03:15.160669 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7rmdd" event={"ID":"59467bee-13bb-4c8c-8617-b485ac0d0de0","Type":"ContainerStarted","Data":"722a80807a9fbfa646c7e0a173ac9b8bfcf496841908a449acdfe1f451b96f6d"} Dec 09 16:03:15 crc kubenswrapper[4894]: I1209 16:03:15.164366 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 16:03:22 crc kubenswrapper[4894]: I1209 16:03:22.235588 4894 generic.go:334] "Generic (PLEG): container finished" podID="59467bee-13bb-4c8c-8617-b485ac0d0de0" containerID="dde39435d23a41d472f494e1c9f060fa4ecda4e9d6708ec0b324154d33476512" exitCode=0 Dec 09 16:03:22 crc kubenswrapper[4894]: I1209 16:03:22.235671 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7rmdd" event={"ID":"59467bee-13bb-4c8c-8617-b485ac0d0de0","Type":"ContainerDied","Data":"dde39435d23a41d472f494e1c9f060fa4ecda4e9d6708ec0b324154d33476512"} Dec 09 16:03:24 crc kubenswrapper[4894]: I1209 16:03:24.255528 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7rmdd" event={"ID":"59467bee-13bb-4c8c-8617-b485ac0d0de0","Type":"ContainerStarted","Data":"ac902c38d30b45aa18cdb7c6d5a8c84b7d0a1e1b748c200b20afecfb0ad26847"} Dec 09 16:03:24 crc kubenswrapper[4894]: I1209 16:03:24.284826 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7rmdd" podStartSLOduration=3.420976643 podStartE2EDuration="11.284797293s" podCreationTimestamp="2025-12-09 16:03:13 +0000 UTC" firstStartedPulling="2025-12-09 16:03:15.164104019 +0000 UTC m=+1889.483314688" lastFinishedPulling="2025-12-09 16:03:23.027924669 +0000 UTC m=+1897.347135338" observedRunningTime="2025-12-09 16:03:24.283300153 +0000 UTC m=+1898.602510842" watchObservedRunningTime="2025-12-09 16:03:24.284797293 +0000 UTC m=+1898.604008002" Dec 09 16:03:34 crc kubenswrapper[4894]: I1209 16:03:34.146689 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7rmdd" Dec 09 16:03:34 crc kubenswrapper[4894]: I1209 16:03:34.147169 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7rmdd" Dec 09 16:03:34 crc kubenswrapper[4894]: I1209 16:03:34.215058 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7rmdd" Dec 09 16:03:34 crc kubenswrapper[4894]: I1209 16:03:34.424873 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7rmdd" Dec 09 16:03:36 crc kubenswrapper[4894]: I1209 16:03:36.229051 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7rmdd"] Dec 09 16:03:36 crc kubenswrapper[4894]: I1209 16:03:36.534769 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9dcgf"] Dec 09 16:03:36 crc kubenswrapper[4894]: I1209 16:03:36.535154 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-9dcgf" podUID="bfcbb982-3316-4cda-b012-0813b7fd4259" containerName="registry-server" containerID="cri-o://1d7593f67faeea7eaf084f0f7c34171f41fdcb92096a5b20b7de1312d1d4c7e6" gracePeriod=2 Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.009077 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9dcgf" Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.024954 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mncp4\" (UniqueName: \"kubernetes.io/projected/bfcbb982-3316-4cda-b012-0813b7fd4259-kube-api-access-mncp4\") pod \"bfcbb982-3316-4cda-b012-0813b7fd4259\" (UID: \"bfcbb982-3316-4cda-b012-0813b7fd4259\") " Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.025059 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfcbb982-3316-4cda-b012-0813b7fd4259-catalog-content\") pod \"bfcbb982-3316-4cda-b012-0813b7fd4259\" (UID: \"bfcbb982-3316-4cda-b012-0813b7fd4259\") " Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.025088 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfcbb982-3316-4cda-b012-0813b7fd4259-utilities\") pod \"bfcbb982-3316-4cda-b012-0813b7fd4259\" (UID: \"bfcbb982-3316-4cda-b012-0813b7fd4259\") " Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.027798 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfcbb982-3316-4cda-b012-0813b7fd4259-utilities" (OuterVolumeSpecName: "utilities") pod "bfcbb982-3316-4cda-b012-0813b7fd4259" (UID: "bfcbb982-3316-4cda-b012-0813b7fd4259"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.037805 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bfcbb982-3316-4cda-b012-0813b7fd4259-kube-api-access-mncp4" (OuterVolumeSpecName: "kube-api-access-mncp4") pod "bfcbb982-3316-4cda-b012-0813b7fd4259" (UID: "bfcbb982-3316-4cda-b012-0813b7fd4259"). InnerVolumeSpecName "kube-api-access-mncp4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.095385 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bfcbb982-3316-4cda-b012-0813b7fd4259-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bfcbb982-3316-4cda-b012-0813b7fd4259" (UID: "bfcbb982-3316-4cda-b012-0813b7fd4259"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.127733 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mncp4\" (UniqueName: \"kubernetes.io/projected/bfcbb982-3316-4cda-b012-0813b7fd4259-kube-api-access-mncp4\") on node \"crc\" DevicePath \"\"" Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.127760 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bfcbb982-3316-4cda-b012-0813b7fd4259-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.127771 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bfcbb982-3316-4cda-b012-0813b7fd4259-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.392151 4894 generic.go:334] "Generic (PLEG): container finished" podID="bfcbb982-3316-4cda-b012-0813b7fd4259" containerID="1d7593f67faeea7eaf084f0f7c34171f41fdcb92096a5b20b7de1312d1d4c7e6" exitCode=0 Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.392205 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9dcgf" event={"ID":"bfcbb982-3316-4cda-b012-0813b7fd4259","Type":"ContainerDied","Data":"1d7593f67faeea7eaf084f0f7c34171f41fdcb92096a5b20b7de1312d1d4c7e6"} Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.392237 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-9dcgf" event={"ID":"bfcbb982-3316-4cda-b012-0813b7fd4259","Type":"ContainerDied","Data":"64cde353590a62be16c6ac0542938adc7bdb6b02c8a628def25c07023579ae7c"} Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.392257 4894 scope.go:117] "RemoveContainer" containerID="1d7593f67faeea7eaf084f0f7c34171f41fdcb92096a5b20b7de1312d1d4c7e6" Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.392411 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-9dcgf" Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.421487 4894 scope.go:117] "RemoveContainer" containerID="89ddccc4bb8b2ff1669dcd7f11f8afd573a55d4a68090eb0af54957223678427" Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.434387 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-9dcgf"] Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.442054 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-9dcgf"] Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.450316 4894 scope.go:117] "RemoveContainer" containerID="a95235df854dd9332191b34d0879a6e97824ae9b9a73e4ccd97151ce443be31b" Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.492413 4894 scope.go:117] "RemoveContainer" containerID="1d7593f67faeea7eaf084f0f7c34171f41fdcb92096a5b20b7de1312d1d4c7e6" Dec 09 16:03:37 crc kubenswrapper[4894]: E1209 16:03:37.492821 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d7593f67faeea7eaf084f0f7c34171f41fdcb92096a5b20b7de1312d1d4c7e6\": container with ID starting with 1d7593f67faeea7eaf084f0f7c34171f41fdcb92096a5b20b7de1312d1d4c7e6 not found: ID does not exist" containerID="1d7593f67faeea7eaf084f0f7c34171f41fdcb92096a5b20b7de1312d1d4c7e6" Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.492850 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d7593f67faeea7eaf084f0f7c34171f41fdcb92096a5b20b7de1312d1d4c7e6"} err="failed to get container status \"1d7593f67faeea7eaf084f0f7c34171f41fdcb92096a5b20b7de1312d1d4c7e6\": rpc error: code = NotFound desc = could not find container \"1d7593f67faeea7eaf084f0f7c34171f41fdcb92096a5b20b7de1312d1d4c7e6\": container with ID starting with 1d7593f67faeea7eaf084f0f7c34171f41fdcb92096a5b20b7de1312d1d4c7e6 not found: ID does not exist" Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.492871 4894 scope.go:117] "RemoveContainer" containerID="89ddccc4bb8b2ff1669dcd7f11f8afd573a55d4a68090eb0af54957223678427" Dec 09 16:03:37 crc kubenswrapper[4894]: E1209 16:03:37.493377 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"89ddccc4bb8b2ff1669dcd7f11f8afd573a55d4a68090eb0af54957223678427\": container with ID starting with 89ddccc4bb8b2ff1669dcd7f11f8afd573a55d4a68090eb0af54957223678427 not found: ID does not exist" containerID="89ddccc4bb8b2ff1669dcd7f11f8afd573a55d4a68090eb0af54957223678427" Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.493418 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"89ddccc4bb8b2ff1669dcd7f11f8afd573a55d4a68090eb0af54957223678427"} err="failed to get container status \"89ddccc4bb8b2ff1669dcd7f11f8afd573a55d4a68090eb0af54957223678427\": rpc error: code = NotFound desc = could not find container \"89ddccc4bb8b2ff1669dcd7f11f8afd573a55d4a68090eb0af54957223678427\": container with ID starting with 89ddccc4bb8b2ff1669dcd7f11f8afd573a55d4a68090eb0af54957223678427 not found: ID does not exist" Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.493443 4894 scope.go:117] "RemoveContainer" containerID="a95235df854dd9332191b34d0879a6e97824ae9b9a73e4ccd97151ce443be31b" Dec 09 16:03:37 crc kubenswrapper[4894]: E1209 16:03:37.493847 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a95235df854dd9332191b34d0879a6e97824ae9b9a73e4ccd97151ce443be31b\": container with ID starting with a95235df854dd9332191b34d0879a6e97824ae9b9a73e4ccd97151ce443be31b not found: ID does not exist" containerID="a95235df854dd9332191b34d0879a6e97824ae9b9a73e4ccd97151ce443be31b" Dec 09 16:03:37 crc kubenswrapper[4894]: I1209 16:03:37.493898 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a95235df854dd9332191b34d0879a6e97824ae9b9a73e4ccd97151ce443be31b"} err="failed to get container status \"a95235df854dd9332191b34d0879a6e97824ae9b9a73e4ccd97151ce443be31b\": rpc error: code = NotFound desc = could not find container \"a95235df854dd9332191b34d0879a6e97824ae9b9a73e4ccd97151ce443be31b\": container with ID starting with a95235df854dd9332191b34d0879a6e97824ae9b9a73e4ccd97151ce443be31b not found: ID does not exist" Dec 09 16:03:38 crc kubenswrapper[4894]: I1209 16:03:38.118321 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bfcbb982-3316-4cda-b012-0813b7fd4259" path="/var/lib/kubelet/pods/bfcbb982-3316-4cda-b012-0813b7fd4259/volumes" Dec 09 16:04:42 crc kubenswrapper[4894]: I1209 16:04:42.257969 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:04:42 crc kubenswrapper[4894]: I1209 16:04:42.258825 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:05:12 crc kubenswrapper[4894]: I1209 16:05:12.257426 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:05:12 crc kubenswrapper[4894]: I1209 16:05:12.258299 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:05:42 crc kubenswrapper[4894]: I1209 16:05:42.257364 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:05:42 crc kubenswrapper[4894]: I1209 16:05:42.258066 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:05:42 crc kubenswrapper[4894]: I1209 16:05:42.258140 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 16:05:42 crc kubenswrapper[4894]: I1209 16:05:42.259259 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"b62d7b69e0742d175c7cec7191590d1fa9926a45adb97c2aa898cb0cc3a77b56"} pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 16:05:42 crc kubenswrapper[4894]: I1209 16:05:42.259378 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" containerID="cri-o://b62d7b69e0742d175c7cec7191590d1fa9926a45adb97c2aa898cb0cc3a77b56" gracePeriod=600 Dec 09 16:05:42 crc kubenswrapper[4894]: I1209 16:05:42.656360 4894 generic.go:334] "Generic (PLEG): container finished" podID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerID="b62d7b69e0742d175c7cec7191590d1fa9926a45adb97c2aa898cb0cc3a77b56" exitCode=0 Dec 09 16:05:42 crc kubenswrapper[4894]: I1209 16:05:42.656436 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerDied","Data":"b62d7b69e0742d175c7cec7191590d1fa9926a45adb97c2aa898cb0cc3a77b56"} Dec 09 16:05:42 crc kubenswrapper[4894]: I1209 16:05:42.656741 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerStarted","Data":"fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678"} Dec 09 16:05:42 crc kubenswrapper[4894]: I1209 16:05:42.656774 4894 scope.go:117] "RemoveContainer" containerID="5ea9681450d298e3dd74dcd8173b2a8ad039507ca18b79c78bc1c9eb3c6d5449" Dec 09 16:06:08 crc kubenswrapper[4894]: I1209 16:06:08.323516 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f"] Dec 09 16:06:08 crc kubenswrapper[4894]: I1209 16:06:08.335198 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-tzfrh"] Dec 09 16:06:08 crc kubenswrapper[4894]: I1209 16:06:08.347533 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-qk49f"] Dec 09 16:06:08 crc kubenswrapper[4894]: I1209 16:06:08.355303 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx"] Dec 09 16:06:08 crc kubenswrapper[4894]: I1209 16:06:08.361913 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p"] Dec 09 16:06:08 crc kubenswrapper[4894]: I1209 16:06:08.368626 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b"] Dec 09 16:06:08 crc kubenswrapper[4894]: I1209 16:06:08.375352 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-mlmfx"] Dec 09 16:06:08 crc kubenswrapper[4894]: I1209 16:06:08.381979 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-97l8p"] Dec 09 16:06:08 crc kubenswrapper[4894]: I1209 16:06:08.388835 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-tzfrh"] Dec 09 16:06:08 crc kubenswrapper[4894]: I1209 16:06:08.395344 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp"] Dec 09 16:06:08 crc kubenswrapper[4894]: I1209 16:06:08.401670 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4"] Dec 09 16:06:08 crc kubenswrapper[4894]: I1209 16:06:08.409491 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n"] Dec 09 16:06:08 crc kubenswrapper[4894]: I1209 16:06:08.416586 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-ntpv4"] Dec 09 16:06:08 crc kubenswrapper[4894]: I1209 16:06:08.423738 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-q4q6n"] Dec 09 16:06:08 crc kubenswrapper[4894]: I1209 16:06:08.430523 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44"] Dec 09 16:06:08 crc kubenswrapper[4894]: I1209 16:06:08.436247 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh"] Dec 09 16:06:08 crc kubenswrapper[4894]: I1209 16:06:08.442067 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-76p8b"] Dec 09 16:06:08 crc kubenswrapper[4894]: I1209 16:06:08.448477 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-pxwsp"] Dec 09 16:06:08 crc kubenswrapper[4894]: I1209 16:06:08.454436 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-jhh44"] Dec 09 16:06:08 crc kubenswrapper[4894]: I1209 16:06:08.461086 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-8tlgh"] Dec 09 16:06:10 crc kubenswrapper[4894]: I1209 16:06:10.117820 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1" path="/var/lib/kubelet/pods/22ddcf3b-9cab-4b3c-aa25-7477e2a7b5b1/volumes" Dec 09 16:06:10 crc kubenswrapper[4894]: I1209 16:06:10.118889 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3577fde1-09d2-456c-aaf8-797e1325775c" path="/var/lib/kubelet/pods/3577fde1-09d2-456c-aaf8-797e1325775c/volumes" Dec 09 16:06:10 crc kubenswrapper[4894]: I1209 16:06:10.119386 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7dec777f-52ed-4e72-8510-352d8da796f7" path="/var/lib/kubelet/pods/7dec777f-52ed-4e72-8510-352d8da796f7/volumes" Dec 09 16:06:10 crc kubenswrapper[4894]: I1209 16:06:10.119923 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2" path="/var/lib/kubelet/pods/829d6c52-d0c4-4e33-b6ed-feb3bc2a74e2/volumes" Dec 09 16:06:10 crc kubenswrapper[4894]: I1209 16:06:10.121229 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="94a3b079-5c34-4ef3-9262-eb741ef7910b" path="/var/lib/kubelet/pods/94a3b079-5c34-4ef3-9262-eb741ef7910b/volumes" Dec 09 16:06:10 crc kubenswrapper[4894]: I1209 16:06:10.121879 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9e6b8027-8cc2-460e-8749-3ceebeee52ad" path="/var/lib/kubelet/pods/9e6b8027-8cc2-460e-8749-3ceebeee52ad/volumes" Dec 09 16:06:10 crc kubenswrapper[4894]: I1209 16:06:10.122795 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b5ef8626-14e1-4fb2-a137-2bf35b64146d" path="/var/lib/kubelet/pods/b5ef8626-14e1-4fb2-a137-2bf35b64146d/volumes" Dec 09 16:06:10 crc kubenswrapper[4894]: I1209 16:06:10.123904 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be481ff5-0866-4a78-84a5-2963c405dd12" path="/var/lib/kubelet/pods/be481ff5-0866-4a78-84a5-2963c405dd12/volumes" Dec 09 16:06:10 crc kubenswrapper[4894]: I1209 16:06:10.124378 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db9fb459-0a0a-4a5f-8510-12adae609c7f" path="/var/lib/kubelet/pods/db9fb459-0a0a-4a5f-8510-12adae609c7f/volumes" Dec 09 16:06:10 crc kubenswrapper[4894]: I1209 16:06:10.125069 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3ac5cab-cda9-48f5-b042-3cc149a120d5" path="/var/lib/kubelet/pods/f3ac5cab-cda9-48f5-b042-3cc149a120d5/volumes" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.089298 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6"] Dec 09 16:06:14 crc kubenswrapper[4894]: E1209 16:06:14.090329 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfcbb982-3316-4cda-b012-0813b7fd4259" containerName="extract-utilities" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.090349 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfcbb982-3316-4cda-b012-0813b7fd4259" containerName="extract-utilities" Dec 09 16:06:14 crc kubenswrapper[4894]: E1209 16:06:14.090361 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfcbb982-3316-4cda-b012-0813b7fd4259" containerName="registry-server" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.090368 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfcbb982-3316-4cda-b012-0813b7fd4259" containerName="registry-server" Dec 09 16:06:14 crc kubenswrapper[4894]: E1209 16:06:14.090380 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bfcbb982-3316-4cda-b012-0813b7fd4259" containerName="extract-content" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.090389 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="bfcbb982-3316-4cda-b012-0813b7fd4259" containerName="extract-content" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.090618 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="bfcbb982-3316-4cda-b012-0813b7fd4259" containerName="registry-server" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.091434 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.103489 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.103725 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6"] Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.104184 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.104511 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.104706 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.104893 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.166207 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6\" (UID: \"066bd81f-b24f-421c-afbc-cc34dfc33828\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.166336 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6\" (UID: \"066bd81f-b24f-421c-afbc-cc34dfc33828\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.166401 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6\" (UID: \"066bd81f-b24f-421c-afbc-cc34dfc33828\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.166462 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6\" (UID: \"066bd81f-b24f-421c-afbc-cc34dfc33828\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.166490 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x2294\" (UniqueName: \"kubernetes.io/projected/066bd81f-b24f-421c-afbc-cc34dfc33828-kube-api-access-x2294\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6\" (UID: \"066bd81f-b24f-421c-afbc-cc34dfc33828\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.268091 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6\" (UID: \"066bd81f-b24f-421c-afbc-cc34dfc33828\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.268210 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6\" (UID: \"066bd81f-b24f-421c-afbc-cc34dfc33828\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.268268 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6\" (UID: \"066bd81f-b24f-421c-afbc-cc34dfc33828\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.268306 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x2294\" (UniqueName: \"kubernetes.io/projected/066bd81f-b24f-421c-afbc-cc34dfc33828-kube-api-access-x2294\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6\" (UID: \"066bd81f-b24f-421c-afbc-cc34dfc33828\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.268376 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6\" (UID: \"066bd81f-b24f-421c-afbc-cc34dfc33828\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.275030 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-inventory\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6\" (UID: \"066bd81f-b24f-421c-afbc-cc34dfc33828\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.275288 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-ceph\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6\" (UID: \"066bd81f-b24f-421c-afbc-cc34dfc33828\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.275380 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-repo-setup-combined-ca-bundle\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6\" (UID: \"066bd81f-b24f-421c-afbc-cc34dfc33828\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.275531 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-ssh-key\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6\" (UID: \"066bd81f-b24f-421c-afbc-cc34dfc33828\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.284741 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x2294\" (UniqueName: \"kubernetes.io/projected/066bd81f-b24f-421c-afbc-cc34dfc33828-kube-api-access-x2294\") pod \"repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6\" (UID: \"066bd81f-b24f-421c-afbc-cc34dfc33828\") " pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.416777 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.806042 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6"] Dec 09 16:06:14 crc kubenswrapper[4894]: W1209 16:06:14.835593 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod066bd81f_b24f_421c_afbc_cc34dfc33828.slice/crio-8aa955e0ff1981430e5070760d1490014c9b5a82fb4eb05df4e3f8b68faccc91 WatchSource:0}: Error finding container 8aa955e0ff1981430e5070760d1490014c9b5a82fb4eb05df4e3f8b68faccc91: Status 404 returned error can't find the container with id 8aa955e0ff1981430e5070760d1490014c9b5a82fb4eb05df4e3f8b68faccc91 Dec 09 16:06:14 crc kubenswrapper[4894]: I1209 16:06:14.959534 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" event={"ID":"066bd81f-b24f-421c-afbc-cc34dfc33828","Type":"ContainerStarted","Data":"8aa955e0ff1981430e5070760d1490014c9b5a82fb4eb05df4e3f8b68faccc91"} Dec 09 16:06:19 crc kubenswrapper[4894]: I1209 16:06:19.000840 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" event={"ID":"066bd81f-b24f-421c-afbc-cc34dfc33828","Type":"ContainerStarted","Data":"d8ec97048e8cd7b71f79799100ef71cc029b56f408bd8425ad577c299de08013"} Dec 09 16:06:19 crc kubenswrapper[4894]: I1209 16:06:19.022823 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" podStartSLOduration=1.715378131 podStartE2EDuration="5.022799476s" podCreationTimestamp="2025-12-09 16:06:14 +0000 UTC" firstStartedPulling="2025-12-09 16:06:14.841210685 +0000 UTC m=+2069.160421354" lastFinishedPulling="2025-12-09 16:06:18.14863203 +0000 UTC m=+2072.467842699" observedRunningTime="2025-12-09 16:06:19.018566592 +0000 UTC m=+2073.337777261" watchObservedRunningTime="2025-12-09 16:06:19.022799476 +0000 UTC m=+2073.342010145" Dec 09 16:06:30 crc kubenswrapper[4894]: I1209 16:06:30.103555 4894 generic.go:334] "Generic (PLEG): container finished" podID="066bd81f-b24f-421c-afbc-cc34dfc33828" containerID="d8ec97048e8cd7b71f79799100ef71cc029b56f408bd8425ad577c299de08013" exitCode=0 Dec 09 16:06:30 crc kubenswrapper[4894]: I1209 16:06:30.103679 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" event={"ID":"066bd81f-b24f-421c-afbc-cc34dfc33828","Type":"ContainerDied","Data":"d8ec97048e8cd7b71f79799100ef71cc029b56f408bd8425ad577c299de08013"} Dec 09 16:06:31 crc kubenswrapper[4894]: I1209 16:06:31.556616 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" Dec 09 16:06:31 crc kubenswrapper[4894]: I1209 16:06:31.744455 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-ssh-key\") pod \"066bd81f-b24f-421c-afbc-cc34dfc33828\" (UID: \"066bd81f-b24f-421c-afbc-cc34dfc33828\") " Dec 09 16:06:31 crc kubenswrapper[4894]: I1209 16:06:31.744506 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-repo-setup-combined-ca-bundle\") pod \"066bd81f-b24f-421c-afbc-cc34dfc33828\" (UID: \"066bd81f-b24f-421c-afbc-cc34dfc33828\") " Dec 09 16:06:31 crc kubenswrapper[4894]: I1209 16:06:31.744628 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2294\" (UniqueName: \"kubernetes.io/projected/066bd81f-b24f-421c-afbc-cc34dfc33828-kube-api-access-x2294\") pod \"066bd81f-b24f-421c-afbc-cc34dfc33828\" (UID: \"066bd81f-b24f-421c-afbc-cc34dfc33828\") " Dec 09 16:06:31 crc kubenswrapper[4894]: I1209 16:06:31.744794 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-ceph\") pod \"066bd81f-b24f-421c-afbc-cc34dfc33828\" (UID: \"066bd81f-b24f-421c-afbc-cc34dfc33828\") " Dec 09 16:06:31 crc kubenswrapper[4894]: I1209 16:06:31.744843 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-inventory\") pod \"066bd81f-b24f-421c-afbc-cc34dfc33828\" (UID: \"066bd81f-b24f-421c-afbc-cc34dfc33828\") " Dec 09 16:06:31 crc kubenswrapper[4894]: I1209 16:06:31.750824 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-ceph" (OuterVolumeSpecName: "ceph") pod "066bd81f-b24f-421c-afbc-cc34dfc33828" (UID: "066bd81f-b24f-421c-afbc-cc34dfc33828"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:06:31 crc kubenswrapper[4894]: I1209 16:06:31.750966 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/066bd81f-b24f-421c-afbc-cc34dfc33828-kube-api-access-x2294" (OuterVolumeSpecName: "kube-api-access-x2294") pod "066bd81f-b24f-421c-afbc-cc34dfc33828" (UID: "066bd81f-b24f-421c-afbc-cc34dfc33828"). InnerVolumeSpecName "kube-api-access-x2294". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:06:31 crc kubenswrapper[4894]: I1209 16:06:31.758890 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "066bd81f-b24f-421c-afbc-cc34dfc33828" (UID: "066bd81f-b24f-421c-afbc-cc34dfc33828"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:06:31 crc kubenswrapper[4894]: I1209 16:06:31.775721 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-inventory" (OuterVolumeSpecName: "inventory") pod "066bd81f-b24f-421c-afbc-cc34dfc33828" (UID: "066bd81f-b24f-421c-afbc-cc34dfc33828"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:06:31 crc kubenswrapper[4894]: I1209 16:06:31.783991 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "066bd81f-b24f-421c-afbc-cc34dfc33828" (UID: "066bd81f-b24f-421c-afbc-cc34dfc33828"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:06:31 crc kubenswrapper[4894]: I1209 16:06:31.847411 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:06:31 crc kubenswrapper[4894]: I1209 16:06:31.847450 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 16:06:31 crc kubenswrapper[4894]: I1209 16:06:31.847463 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:06:31 crc kubenswrapper[4894]: I1209 16:06:31.847475 4894 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/066bd81f-b24f-421c-afbc-cc34dfc33828-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:06:31 crc kubenswrapper[4894]: I1209 16:06:31.847487 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2294\" (UniqueName: \"kubernetes.io/projected/066bd81f-b24f-421c-afbc-cc34dfc33828-kube-api-access-x2294\") on node \"crc\" DevicePath \"\"" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.135750 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" event={"ID":"066bd81f-b24f-421c-afbc-cc34dfc33828","Type":"ContainerDied","Data":"8aa955e0ff1981430e5070760d1490014c9b5a82fb4eb05df4e3f8b68faccc91"} Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.135798 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8aa955e0ff1981430e5070760d1490014c9b5a82fb4eb05df4e3f8b68faccc91" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.135824 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.226197 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv"] Dec 09 16:06:32 crc kubenswrapper[4894]: E1209 16:06:32.226973 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="066bd81f-b24f-421c-afbc-cc34dfc33828" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.227041 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="066bd81f-b24f-421c-afbc-cc34dfc33828" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.227348 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="066bd81f-b24f-421c-afbc-cc34dfc33828" containerName="repo-setup-edpm-deployment-openstack-edpm-ipam" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.228433 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.230251 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.232859 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.232932 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.232960 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.233046 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.236409 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv"] Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.256215 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nksff\" (UniqueName: \"kubernetes.io/projected/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-kube-api-access-nksff\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv\" (UID: \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.256281 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv\" (UID: \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.256398 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv\" (UID: \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.256486 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv\" (UID: \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.256517 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv\" (UID: \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.357329 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv\" (UID: \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.357368 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv\" (UID: \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.357433 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nksff\" (UniqueName: \"kubernetes.io/projected/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-kube-api-access-nksff\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv\" (UID: \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.357470 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv\" (UID: \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.357514 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv\" (UID: \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.361274 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-ssh-key\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv\" (UID: \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.361282 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-inventory\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv\" (UID: \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.361944 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-ceph\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv\" (UID: \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.364190 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-bootstrap-combined-ca-bundle\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv\" (UID: \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.374552 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nksff\" (UniqueName: \"kubernetes.io/projected/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-kube-api-access-nksff\") pod \"bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv\" (UID: \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\") " pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" Dec 09 16:06:32 crc kubenswrapper[4894]: I1209 16:06:32.555187 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" Dec 09 16:06:33 crc kubenswrapper[4894]: I1209 16:06:33.147409 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv"] Dec 09 16:06:34 crc kubenswrapper[4894]: I1209 16:06:34.165338 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" event={"ID":"6f620852-03c1-48ea-ae2d-b0f5de4ab16b","Type":"ContainerStarted","Data":"f8b03334a8d707eec0fddbcb124d94fef3b46abaf797d6beb2b60e4e3aafd2ee"} Dec 09 16:06:36 crc kubenswrapper[4894]: I1209 16:06:36.197717 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" event={"ID":"6f620852-03c1-48ea-ae2d-b0f5de4ab16b","Type":"ContainerStarted","Data":"14d66fa17b9f38cc898574b1d1b540ac5751ec46be67bb51ea0f53b094195d7f"} Dec 09 16:06:37 crc kubenswrapper[4894]: I1209 16:06:37.235555 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" podStartSLOduration=3.210298375 podStartE2EDuration="5.235536933s" podCreationTimestamp="2025-12-09 16:06:32 +0000 UTC" firstStartedPulling="2025-12-09 16:06:33.162322224 +0000 UTC m=+2087.481532893" lastFinishedPulling="2025-12-09 16:06:35.187560762 +0000 UTC m=+2089.506771451" observedRunningTime="2025-12-09 16:06:37.232035388 +0000 UTC m=+2091.551246057" watchObservedRunningTime="2025-12-09 16:06:37.235536933 +0000 UTC m=+2091.554747602" Dec 09 16:06:39 crc kubenswrapper[4894]: I1209 16:06:39.127131 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-jhjxg"] Dec 09 16:06:39 crc kubenswrapper[4894]: I1209 16:06:39.128771 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jhjxg" Dec 09 16:06:39 crc kubenswrapper[4894]: I1209 16:06:39.148133 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jhjxg"] Dec 09 16:06:39 crc kubenswrapper[4894]: I1209 16:06:39.292014 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdmd2\" (UniqueName: \"kubernetes.io/projected/2a3d33af-2799-4c66-b963-3d6c70036854-kube-api-access-xdmd2\") pod \"redhat-operators-jhjxg\" (UID: \"2a3d33af-2799-4c66-b963-3d6c70036854\") " pod="openshift-marketplace/redhat-operators-jhjxg" Dec 09 16:06:39 crc kubenswrapper[4894]: I1209 16:06:39.292428 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a3d33af-2799-4c66-b963-3d6c70036854-catalog-content\") pod \"redhat-operators-jhjxg\" (UID: \"2a3d33af-2799-4c66-b963-3d6c70036854\") " pod="openshift-marketplace/redhat-operators-jhjxg" Dec 09 16:06:39 crc kubenswrapper[4894]: I1209 16:06:39.292482 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a3d33af-2799-4c66-b963-3d6c70036854-utilities\") pod \"redhat-operators-jhjxg\" (UID: \"2a3d33af-2799-4c66-b963-3d6c70036854\") " pod="openshift-marketplace/redhat-operators-jhjxg" Dec 09 16:06:39 crc kubenswrapper[4894]: I1209 16:06:39.394695 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdmd2\" (UniqueName: \"kubernetes.io/projected/2a3d33af-2799-4c66-b963-3d6c70036854-kube-api-access-xdmd2\") pod \"redhat-operators-jhjxg\" (UID: \"2a3d33af-2799-4c66-b963-3d6c70036854\") " pod="openshift-marketplace/redhat-operators-jhjxg" Dec 09 16:06:39 crc kubenswrapper[4894]: I1209 16:06:39.394858 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a3d33af-2799-4c66-b963-3d6c70036854-catalog-content\") pod \"redhat-operators-jhjxg\" (UID: \"2a3d33af-2799-4c66-b963-3d6c70036854\") " pod="openshift-marketplace/redhat-operators-jhjxg" Dec 09 16:06:39 crc kubenswrapper[4894]: I1209 16:06:39.394908 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a3d33af-2799-4c66-b963-3d6c70036854-utilities\") pod \"redhat-operators-jhjxg\" (UID: \"2a3d33af-2799-4c66-b963-3d6c70036854\") " pod="openshift-marketplace/redhat-operators-jhjxg" Dec 09 16:06:39 crc kubenswrapper[4894]: I1209 16:06:39.395416 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a3d33af-2799-4c66-b963-3d6c70036854-catalog-content\") pod \"redhat-operators-jhjxg\" (UID: \"2a3d33af-2799-4c66-b963-3d6c70036854\") " pod="openshift-marketplace/redhat-operators-jhjxg" Dec 09 16:06:39 crc kubenswrapper[4894]: I1209 16:06:39.395439 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a3d33af-2799-4c66-b963-3d6c70036854-utilities\") pod \"redhat-operators-jhjxg\" (UID: \"2a3d33af-2799-4c66-b963-3d6c70036854\") " pod="openshift-marketplace/redhat-operators-jhjxg" Dec 09 16:06:39 crc kubenswrapper[4894]: I1209 16:06:39.416730 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdmd2\" (UniqueName: \"kubernetes.io/projected/2a3d33af-2799-4c66-b963-3d6c70036854-kube-api-access-xdmd2\") pod \"redhat-operators-jhjxg\" (UID: \"2a3d33af-2799-4c66-b963-3d6c70036854\") " pod="openshift-marketplace/redhat-operators-jhjxg" Dec 09 16:06:39 crc kubenswrapper[4894]: I1209 16:06:39.462485 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jhjxg" Dec 09 16:06:39 crc kubenswrapper[4894]: I1209 16:06:39.930379 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jhjxg"] Dec 09 16:06:40 crc kubenswrapper[4894]: I1209 16:06:40.231747 4894 generic.go:334] "Generic (PLEG): container finished" podID="2a3d33af-2799-4c66-b963-3d6c70036854" containerID="0bb033a956a7bc057898f56b56faf95ed01c561ce6d3f07138ee162824bb5def" exitCode=0 Dec 09 16:06:40 crc kubenswrapper[4894]: I1209 16:06:40.231839 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhjxg" event={"ID":"2a3d33af-2799-4c66-b963-3d6c70036854","Type":"ContainerDied","Data":"0bb033a956a7bc057898f56b56faf95ed01c561ce6d3f07138ee162824bb5def"} Dec 09 16:06:40 crc kubenswrapper[4894]: I1209 16:06:40.232034 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhjxg" event={"ID":"2a3d33af-2799-4c66-b963-3d6c70036854","Type":"ContainerStarted","Data":"c5b042877a52da91891ce8e43cc987fa04ce95c8c06441d386033865cd594216"} Dec 09 16:06:51 crc kubenswrapper[4894]: I1209 16:06:51.331717 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhjxg" event={"ID":"2a3d33af-2799-4c66-b963-3d6c70036854","Type":"ContainerStarted","Data":"b8608fd119dd49bd3b5167b25d661a2c0a801cad26384f2f3c165a0ab4dbce88"} Dec 09 16:06:52 crc kubenswrapper[4894]: I1209 16:06:52.341317 4894 generic.go:334] "Generic (PLEG): container finished" podID="2a3d33af-2799-4c66-b963-3d6c70036854" containerID="b8608fd119dd49bd3b5167b25d661a2c0a801cad26384f2f3c165a0ab4dbce88" exitCode=0 Dec 09 16:06:52 crc kubenswrapper[4894]: I1209 16:06:52.341385 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhjxg" event={"ID":"2a3d33af-2799-4c66-b963-3d6c70036854","Type":"ContainerDied","Data":"b8608fd119dd49bd3b5167b25d661a2c0a801cad26384f2f3c165a0ab4dbce88"} Dec 09 16:06:54 crc kubenswrapper[4894]: I1209 16:06:54.359543 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhjxg" event={"ID":"2a3d33af-2799-4c66-b963-3d6c70036854","Type":"ContainerStarted","Data":"f9efed26ba05df2e8a855d216e6a2aa579d27b9c00cedb9a568d118b94aee309"} Dec 09 16:06:54 crc kubenswrapper[4894]: I1209 16:06:54.384523 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-jhjxg" podStartSLOduration=2.400391633 podStartE2EDuration="15.384505701s" podCreationTimestamp="2025-12-09 16:06:39 +0000 UTC" firstStartedPulling="2025-12-09 16:06:40.234867192 +0000 UTC m=+2094.554077861" lastFinishedPulling="2025-12-09 16:06:53.21898126 +0000 UTC m=+2107.538191929" observedRunningTime="2025-12-09 16:06:54.380857263 +0000 UTC m=+2108.700067932" watchObservedRunningTime="2025-12-09 16:06:54.384505701 +0000 UTC m=+2108.703716370" Dec 09 16:06:56 crc kubenswrapper[4894]: I1209 16:06:56.544871 4894 scope.go:117] "RemoveContainer" containerID="6df01dbe7ee500afeee4fdf4e66c4836d9a37ef755c3d2a40fc7b6646637f264" Dec 09 16:06:56 crc kubenswrapper[4894]: I1209 16:06:56.612113 4894 scope.go:117] "RemoveContainer" containerID="ad356faf9f26c46787e5421cec638ad9dee018928d9909ef8762fe781d29ed10" Dec 09 16:06:56 crc kubenswrapper[4894]: I1209 16:06:56.667602 4894 scope.go:117] "RemoveContainer" containerID="55c85647833560d5e7e7e230a2b12b9f0ed5bce0cb39436f7e3ea135e2061620" Dec 09 16:06:56 crc kubenswrapper[4894]: I1209 16:06:56.709979 4894 scope.go:117] "RemoveContainer" containerID="e3775888aaa02850b1ef13192ab58c184bf78ba07afa819acff12c67fd4138b3" Dec 09 16:06:56 crc kubenswrapper[4894]: I1209 16:06:56.764922 4894 scope.go:117] "RemoveContainer" containerID="21f39c8734c83c789b43b891077431c05013a3a2112b43c571a86d62a05ee2a2" Dec 09 16:06:56 crc kubenswrapper[4894]: I1209 16:06:56.810260 4894 scope.go:117] "RemoveContainer" containerID="50066c981c62fafecfce155649674f54334eb95319b66113df81cb777e116108" Dec 09 16:06:56 crc kubenswrapper[4894]: I1209 16:06:56.846357 4894 scope.go:117] "RemoveContainer" containerID="ee1242f966f9341754416dc1c1559f4b62f133d92681eccbec0c9269d274d8e5" Dec 09 16:06:59 crc kubenswrapper[4894]: I1209 16:06:59.463538 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-jhjxg" Dec 09 16:06:59 crc kubenswrapper[4894]: I1209 16:06:59.464325 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-jhjxg" Dec 09 16:06:59 crc kubenswrapper[4894]: I1209 16:06:59.509746 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-jhjxg" Dec 09 16:07:00 crc kubenswrapper[4894]: I1209 16:07:00.470994 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-jhjxg" Dec 09 16:07:00 crc kubenswrapper[4894]: I1209 16:07:00.541961 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-jhjxg"] Dec 09 16:07:00 crc kubenswrapper[4894]: I1209 16:07:00.583597 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8hxbx"] Dec 09 16:07:00 crc kubenswrapper[4894]: I1209 16:07:00.584112 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8hxbx" podUID="83805fb6-3997-4aba-a8ca-15c24e4afe4c" containerName="registry-server" containerID="cri-o://f896b45b40fb7bc155a8021c33fa428209e6e4658fa88b152ff574a1141d6e9a" gracePeriod=2 Dec 09 16:07:01 crc kubenswrapper[4894]: I1209 16:07:01.418540 4894 generic.go:334] "Generic (PLEG): container finished" podID="83805fb6-3997-4aba-a8ca-15c24e4afe4c" containerID="f896b45b40fb7bc155a8021c33fa428209e6e4658fa88b152ff574a1141d6e9a" exitCode=0 Dec 09 16:07:01 crc kubenswrapper[4894]: I1209 16:07:01.418633 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8hxbx" event={"ID":"83805fb6-3997-4aba-a8ca-15c24e4afe4c","Type":"ContainerDied","Data":"f896b45b40fb7bc155a8021c33fa428209e6e4658fa88b152ff574a1141d6e9a"} Dec 09 16:07:01 crc kubenswrapper[4894]: I1209 16:07:01.678022 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8hxbx" Dec 09 16:07:01 crc kubenswrapper[4894]: I1209 16:07:01.838593 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mgwhb\" (UniqueName: \"kubernetes.io/projected/83805fb6-3997-4aba-a8ca-15c24e4afe4c-kube-api-access-mgwhb\") pod \"83805fb6-3997-4aba-a8ca-15c24e4afe4c\" (UID: \"83805fb6-3997-4aba-a8ca-15c24e4afe4c\") " Dec 09 16:07:01 crc kubenswrapper[4894]: I1209 16:07:01.838988 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83805fb6-3997-4aba-a8ca-15c24e4afe4c-catalog-content\") pod \"83805fb6-3997-4aba-a8ca-15c24e4afe4c\" (UID: \"83805fb6-3997-4aba-a8ca-15c24e4afe4c\") " Dec 09 16:07:01 crc kubenswrapper[4894]: I1209 16:07:01.839098 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83805fb6-3997-4aba-a8ca-15c24e4afe4c-utilities\") pod \"83805fb6-3997-4aba-a8ca-15c24e4afe4c\" (UID: \"83805fb6-3997-4aba-a8ca-15c24e4afe4c\") " Dec 09 16:07:01 crc kubenswrapper[4894]: I1209 16:07:01.839580 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83805fb6-3997-4aba-a8ca-15c24e4afe4c-utilities" (OuterVolumeSpecName: "utilities") pod "83805fb6-3997-4aba-a8ca-15c24e4afe4c" (UID: "83805fb6-3997-4aba-a8ca-15c24e4afe4c"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:07:01 crc kubenswrapper[4894]: I1209 16:07:01.861817 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83805fb6-3997-4aba-a8ca-15c24e4afe4c-kube-api-access-mgwhb" (OuterVolumeSpecName: "kube-api-access-mgwhb") pod "83805fb6-3997-4aba-a8ca-15c24e4afe4c" (UID: "83805fb6-3997-4aba-a8ca-15c24e4afe4c"). InnerVolumeSpecName "kube-api-access-mgwhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:07:01 crc kubenswrapper[4894]: I1209 16:07:01.934115 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83805fb6-3997-4aba-a8ca-15c24e4afe4c-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "83805fb6-3997-4aba-a8ca-15c24e4afe4c" (UID: "83805fb6-3997-4aba-a8ca-15c24e4afe4c"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:07:01 crc kubenswrapper[4894]: I1209 16:07:01.941648 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/83805fb6-3997-4aba-a8ca-15c24e4afe4c-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:07:01 crc kubenswrapper[4894]: I1209 16:07:01.941679 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/83805fb6-3997-4aba-a8ca-15c24e4afe4c-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:07:01 crc kubenswrapper[4894]: I1209 16:07:01.941690 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mgwhb\" (UniqueName: \"kubernetes.io/projected/83805fb6-3997-4aba-a8ca-15c24e4afe4c-kube-api-access-mgwhb\") on node \"crc\" DevicePath \"\"" Dec 09 16:07:02 crc kubenswrapper[4894]: I1209 16:07:02.431282 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8hxbx" Dec 09 16:07:02 crc kubenswrapper[4894]: I1209 16:07:02.432074 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8hxbx" event={"ID":"83805fb6-3997-4aba-a8ca-15c24e4afe4c","Type":"ContainerDied","Data":"5b12050d435e6d73449e1352c9a5253fd3b12544e1678d44e241649aaa2356da"} Dec 09 16:07:02 crc kubenswrapper[4894]: I1209 16:07:02.432112 4894 scope.go:117] "RemoveContainer" containerID="f896b45b40fb7bc155a8021c33fa428209e6e4658fa88b152ff574a1141d6e9a" Dec 09 16:07:02 crc kubenswrapper[4894]: I1209 16:07:02.451781 4894 scope.go:117] "RemoveContainer" containerID="7729c3bbe467731935f01097c7facf6bc0e2557877d7ba63489131930e381403" Dec 09 16:07:02 crc kubenswrapper[4894]: I1209 16:07:02.455585 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8hxbx"] Dec 09 16:07:02 crc kubenswrapper[4894]: I1209 16:07:02.463350 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8hxbx"] Dec 09 16:07:02 crc kubenswrapper[4894]: I1209 16:07:02.472490 4894 scope.go:117] "RemoveContainer" containerID="17050e0f7ad4a2484af86173376dc1059f8c6ec77c689d9b8411f81670f51d57" Dec 09 16:07:04 crc kubenswrapper[4894]: I1209 16:07:04.134426 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83805fb6-3997-4aba-a8ca-15c24e4afe4c" path="/var/lib/kubelet/pods/83805fb6-3997-4aba-a8ca-15c24e4afe4c/volumes" Dec 09 16:07:42 crc kubenswrapper[4894]: I1209 16:07:42.257699 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:07:42 crc kubenswrapper[4894]: I1209 16:07:42.258240 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:07:57 crc kubenswrapper[4894]: I1209 16:07:57.003992 4894 scope.go:117] "RemoveContainer" containerID="7cbc03646abf25ca6598e20f165df3e84d297fda8fcce968ff2ba15bda949c9a" Dec 09 16:07:57 crc kubenswrapper[4894]: I1209 16:07:57.036109 4894 scope.go:117] "RemoveContainer" containerID="18e7396abc35ba282dc17c5c9603c81460f6a84975fcfede4834c7f155539f72" Dec 09 16:07:57 crc kubenswrapper[4894]: I1209 16:07:57.072999 4894 scope.go:117] "RemoveContainer" containerID="932d21281cf6b4c3932dafbd6482a4f38deb9a110fe81c7256e28761b3dc7591" Dec 09 16:08:12 crc kubenswrapper[4894]: I1209 16:08:12.257851 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:08:12 crc kubenswrapper[4894]: I1209 16:08:12.258389 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:08:18 crc kubenswrapper[4894]: I1209 16:08:18.128937 4894 generic.go:334] "Generic (PLEG): container finished" podID="6f620852-03c1-48ea-ae2d-b0f5de4ab16b" containerID="14d66fa17b9f38cc898574b1d1b540ac5751ec46be67bb51ea0f53b094195d7f" exitCode=0 Dec 09 16:08:18 crc kubenswrapper[4894]: I1209 16:08:18.129054 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" event={"ID":"6f620852-03c1-48ea-ae2d-b0f5de4ab16b","Type":"ContainerDied","Data":"14d66fa17b9f38cc898574b1d1b540ac5751ec46be67bb51ea0f53b094195d7f"} Dec 09 16:08:19 crc kubenswrapper[4894]: I1209 16:08:19.600697 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" Dec 09 16:08:19 crc kubenswrapper[4894]: I1209 16:08:19.666124 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-bootstrap-combined-ca-bundle\") pod \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\" (UID: \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\") " Dec 09 16:08:19 crc kubenswrapper[4894]: I1209 16:08:19.666220 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-ceph\") pod \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\" (UID: \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\") " Dec 09 16:08:19 crc kubenswrapper[4894]: I1209 16:08:19.666248 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-ssh-key\") pod \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\" (UID: \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\") " Dec 09 16:08:19 crc kubenswrapper[4894]: I1209 16:08:19.666284 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-inventory\") pod \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\" (UID: \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\") " Dec 09 16:08:19 crc kubenswrapper[4894]: I1209 16:08:19.666409 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nksff\" (UniqueName: \"kubernetes.io/projected/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-kube-api-access-nksff\") pod \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\" (UID: \"6f620852-03c1-48ea-ae2d-b0f5de4ab16b\") " Dec 09 16:08:19 crc kubenswrapper[4894]: I1209 16:08:19.672795 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-ceph" (OuterVolumeSpecName: "ceph") pod "6f620852-03c1-48ea-ae2d-b0f5de4ab16b" (UID: "6f620852-03c1-48ea-ae2d-b0f5de4ab16b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:08:19 crc kubenswrapper[4894]: I1209 16:08:19.673623 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "6f620852-03c1-48ea-ae2d-b0f5de4ab16b" (UID: "6f620852-03c1-48ea-ae2d-b0f5de4ab16b"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:08:19 crc kubenswrapper[4894]: I1209 16:08:19.695170 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "6f620852-03c1-48ea-ae2d-b0f5de4ab16b" (UID: "6f620852-03c1-48ea-ae2d-b0f5de4ab16b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:08:19 crc kubenswrapper[4894]: I1209 16:08:19.698538 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-kube-api-access-nksff" (OuterVolumeSpecName: "kube-api-access-nksff") pod "6f620852-03c1-48ea-ae2d-b0f5de4ab16b" (UID: "6f620852-03c1-48ea-ae2d-b0f5de4ab16b"). InnerVolumeSpecName "kube-api-access-nksff". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:08:19 crc kubenswrapper[4894]: I1209 16:08:19.716778 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-inventory" (OuterVolumeSpecName: "inventory") pod "6f620852-03c1-48ea-ae2d-b0f5de4ab16b" (UID: "6f620852-03c1-48ea-ae2d-b0f5de4ab16b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:08:19 crc kubenswrapper[4894]: I1209 16:08:19.769038 4894 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:08:19 crc kubenswrapper[4894]: I1209 16:08:19.769072 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:08:19 crc kubenswrapper[4894]: I1209 16:08:19.769081 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:08:19 crc kubenswrapper[4894]: I1209 16:08:19.769094 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 16:08:19 crc kubenswrapper[4894]: I1209 16:08:19.769104 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nksff\" (UniqueName: \"kubernetes.io/projected/6f620852-03c1-48ea-ae2d-b0f5de4ab16b-kube-api-access-nksff\") on node \"crc\" DevicePath \"\"" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.151059 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" event={"ID":"6f620852-03c1-48ea-ae2d-b0f5de4ab16b","Type":"ContainerDied","Data":"f8b03334a8d707eec0fddbcb124d94fef3b46abaf797d6beb2b60e4e3aafd2ee"} Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.151424 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f8b03334a8d707eec0fddbcb124d94fef3b46abaf797d6beb2b60e4e3aafd2ee" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.151278 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.227591 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l"] Dec 09 16:08:20 crc kubenswrapper[4894]: E1209 16:08:20.228409 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83805fb6-3997-4aba-a8ca-15c24e4afe4c" containerName="registry-server" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.228611 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="83805fb6-3997-4aba-a8ca-15c24e4afe4c" containerName="registry-server" Dec 09 16:08:20 crc kubenswrapper[4894]: E1209 16:08:20.228749 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83805fb6-3997-4aba-a8ca-15c24e4afe4c" containerName="extract-utilities" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.228824 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="83805fb6-3997-4aba-a8ca-15c24e4afe4c" containerName="extract-utilities" Dec 09 16:08:20 crc kubenswrapper[4894]: E1209 16:08:20.228900 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f620852-03c1-48ea-ae2d-b0f5de4ab16b" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.228980 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f620852-03c1-48ea-ae2d-b0f5de4ab16b" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 09 16:08:20 crc kubenswrapper[4894]: E1209 16:08:20.229126 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83805fb6-3997-4aba-a8ca-15c24e4afe4c" containerName="extract-content" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.229205 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="83805fb6-3997-4aba-a8ca-15c24e4afe4c" containerName="extract-content" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.229470 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f620852-03c1-48ea-ae2d-b0f5de4ab16b" containerName="bootstrap-edpm-deployment-openstack-edpm-ipam" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.229563 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="83805fb6-3997-4aba-a8ca-15c24e4afe4c" containerName="registry-server" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.230728 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.237731 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.238027 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.238275 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.238417 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.238878 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.241450 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l"] Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.280955 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l\" (UID: \"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.281021 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vvp48\" (UniqueName: \"kubernetes.io/projected/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-kube-api-access-vvp48\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l\" (UID: \"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.281108 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l\" (UID: \"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.281154 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l\" (UID: \"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.383216 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l\" (UID: \"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.383267 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vvp48\" (UniqueName: \"kubernetes.io/projected/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-kube-api-access-vvp48\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l\" (UID: \"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.383333 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l\" (UID: \"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.383366 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l\" (UID: \"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.387864 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-ssh-key\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l\" (UID: \"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.388495 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-inventory\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l\" (UID: \"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.389913 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-ceph\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l\" (UID: \"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.402827 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vvp48\" (UniqueName: \"kubernetes.io/projected/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-kube-api-access-vvp48\") pod \"configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l\" (UID: \"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48\") " pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l" Dec 09 16:08:20 crc kubenswrapper[4894]: I1209 16:08:20.559582 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l" Dec 09 16:08:21 crc kubenswrapper[4894]: I1209 16:08:21.124462 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l"] Dec 09 16:08:21 crc kubenswrapper[4894]: I1209 16:08:21.138668 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 16:08:21 crc kubenswrapper[4894]: I1209 16:08:21.159076 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l" event={"ID":"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48","Type":"ContainerStarted","Data":"df6abb6914f00efc776595cbc32ea24464012968d7a9055d96c57223348b2ebf"} Dec 09 16:08:22 crc kubenswrapper[4894]: I1209 16:08:22.173003 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l" event={"ID":"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48","Type":"ContainerStarted","Data":"363465de5b17cab3325aecb1090dcefe6b9a92cb2ff857edf945cd093accae78"} Dec 09 16:08:22 crc kubenswrapper[4894]: I1209 16:08:22.198226 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l" podStartSLOduration=1.7002117270000001 podStartE2EDuration="2.198204126s" podCreationTimestamp="2025-12-09 16:08:20 +0000 UTC" firstStartedPulling="2025-12-09 16:08:21.138431305 +0000 UTC m=+2195.457641974" lastFinishedPulling="2025-12-09 16:08:21.636423704 +0000 UTC m=+2195.955634373" observedRunningTime="2025-12-09 16:08:22.18875365 +0000 UTC m=+2196.507964339" watchObservedRunningTime="2025-12-09 16:08:22.198204126 +0000 UTC m=+2196.517414805" Dec 09 16:08:41 crc kubenswrapper[4894]: I1209 16:08:41.817882 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-64fzw"] Dec 09 16:08:41 crc kubenswrapper[4894]: I1209 16:08:41.821379 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-64fzw" Dec 09 16:08:41 crc kubenswrapper[4894]: I1209 16:08:41.834947 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-64fzw"] Dec 09 16:08:42 crc kubenswrapper[4894]: I1209 16:08:42.001959 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca3465d2-9009-4eb8-ab31-57af05272571-utilities\") pod \"redhat-marketplace-64fzw\" (UID: \"ca3465d2-9009-4eb8-ab31-57af05272571\") " pod="openshift-marketplace/redhat-marketplace-64fzw" Dec 09 16:08:42 crc kubenswrapper[4894]: I1209 16:08:42.002026 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca3465d2-9009-4eb8-ab31-57af05272571-catalog-content\") pod \"redhat-marketplace-64fzw\" (UID: \"ca3465d2-9009-4eb8-ab31-57af05272571\") " pod="openshift-marketplace/redhat-marketplace-64fzw" Dec 09 16:08:42 crc kubenswrapper[4894]: I1209 16:08:42.002068 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4sd7\" (UniqueName: \"kubernetes.io/projected/ca3465d2-9009-4eb8-ab31-57af05272571-kube-api-access-w4sd7\") pod \"redhat-marketplace-64fzw\" (UID: \"ca3465d2-9009-4eb8-ab31-57af05272571\") " pod="openshift-marketplace/redhat-marketplace-64fzw" Dec 09 16:08:42 crc kubenswrapper[4894]: I1209 16:08:42.104445 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4sd7\" (UniqueName: \"kubernetes.io/projected/ca3465d2-9009-4eb8-ab31-57af05272571-kube-api-access-w4sd7\") pod \"redhat-marketplace-64fzw\" (UID: \"ca3465d2-9009-4eb8-ab31-57af05272571\") " pod="openshift-marketplace/redhat-marketplace-64fzw" Dec 09 16:08:42 crc kubenswrapper[4894]: I1209 16:08:42.104746 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca3465d2-9009-4eb8-ab31-57af05272571-utilities\") pod \"redhat-marketplace-64fzw\" (UID: \"ca3465d2-9009-4eb8-ab31-57af05272571\") " pod="openshift-marketplace/redhat-marketplace-64fzw" Dec 09 16:08:42 crc kubenswrapper[4894]: I1209 16:08:42.104832 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca3465d2-9009-4eb8-ab31-57af05272571-catalog-content\") pod \"redhat-marketplace-64fzw\" (UID: \"ca3465d2-9009-4eb8-ab31-57af05272571\") " pod="openshift-marketplace/redhat-marketplace-64fzw" Dec 09 16:08:42 crc kubenswrapper[4894]: I1209 16:08:42.105329 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca3465d2-9009-4eb8-ab31-57af05272571-utilities\") pod \"redhat-marketplace-64fzw\" (UID: \"ca3465d2-9009-4eb8-ab31-57af05272571\") " pod="openshift-marketplace/redhat-marketplace-64fzw" Dec 09 16:08:42 crc kubenswrapper[4894]: I1209 16:08:42.105507 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca3465d2-9009-4eb8-ab31-57af05272571-catalog-content\") pod \"redhat-marketplace-64fzw\" (UID: \"ca3465d2-9009-4eb8-ab31-57af05272571\") " pod="openshift-marketplace/redhat-marketplace-64fzw" Dec 09 16:08:42 crc kubenswrapper[4894]: I1209 16:08:42.135561 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4sd7\" (UniqueName: \"kubernetes.io/projected/ca3465d2-9009-4eb8-ab31-57af05272571-kube-api-access-w4sd7\") pod \"redhat-marketplace-64fzw\" (UID: \"ca3465d2-9009-4eb8-ab31-57af05272571\") " pod="openshift-marketplace/redhat-marketplace-64fzw" Dec 09 16:08:42 crc kubenswrapper[4894]: I1209 16:08:42.147131 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-64fzw" Dec 09 16:08:42 crc kubenswrapper[4894]: I1209 16:08:42.257895 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:08:42 crc kubenswrapper[4894]: I1209 16:08:42.258182 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:08:42 crc kubenswrapper[4894]: I1209 16:08:42.258224 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 16:08:42 crc kubenswrapper[4894]: I1209 16:08:42.258926 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678"} pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 16:08:42 crc kubenswrapper[4894]: I1209 16:08:42.258972 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" containerID="cri-o://fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" gracePeriod=600 Dec 09 16:08:42 crc kubenswrapper[4894]: I1209 16:08:42.459918 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-64fzw"] Dec 09 16:08:42 crc kubenswrapper[4894]: E1209 16:08:42.942329 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:08:43 crc kubenswrapper[4894]: I1209 16:08:43.356820 4894 generic.go:334] "Generic (PLEG): container finished" podID="ca3465d2-9009-4eb8-ab31-57af05272571" containerID="6df924c4219dbb8957d31f4b29c815434b186b9ea684eb94b04c7c852505a104" exitCode=0 Dec 09 16:08:43 crc kubenswrapper[4894]: I1209 16:08:43.356876 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-64fzw" event={"ID":"ca3465d2-9009-4eb8-ab31-57af05272571","Type":"ContainerDied","Data":"6df924c4219dbb8957d31f4b29c815434b186b9ea684eb94b04c7c852505a104"} Dec 09 16:08:43 crc kubenswrapper[4894]: I1209 16:08:43.357171 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-64fzw" event={"ID":"ca3465d2-9009-4eb8-ab31-57af05272571","Type":"ContainerStarted","Data":"5b4f7eeb0cc1ede9dfbcbb5159ebdedcbadf7ef5deb884e97a339237b363f11e"} Dec 09 16:08:43 crc kubenswrapper[4894]: I1209 16:08:43.359925 4894 generic.go:334] "Generic (PLEG): container finished" podID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" exitCode=0 Dec 09 16:08:43 crc kubenswrapper[4894]: I1209 16:08:43.359984 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerDied","Data":"fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678"} Dec 09 16:08:43 crc kubenswrapper[4894]: I1209 16:08:43.360025 4894 scope.go:117] "RemoveContainer" containerID="b62d7b69e0742d175c7cec7191590d1fa9926a45adb97c2aa898cb0cc3a77b56" Dec 09 16:08:43 crc kubenswrapper[4894]: I1209 16:08:43.360473 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:08:43 crc kubenswrapper[4894]: E1209 16:08:43.360721 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:08:45 crc kubenswrapper[4894]: I1209 16:08:45.382784 4894 generic.go:334] "Generic (PLEG): container finished" podID="ca3465d2-9009-4eb8-ab31-57af05272571" containerID="3cd6d564db6a03c6f24683b35bb3ca0930d2160a0a86c516ce9862494ba3b1bf" exitCode=0 Dec 09 16:08:45 crc kubenswrapper[4894]: I1209 16:08:45.383443 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-64fzw" event={"ID":"ca3465d2-9009-4eb8-ab31-57af05272571","Type":"ContainerDied","Data":"3cd6d564db6a03c6f24683b35bb3ca0930d2160a0a86c516ce9862494ba3b1bf"} Dec 09 16:08:47 crc kubenswrapper[4894]: I1209 16:08:47.408627 4894 generic.go:334] "Generic (PLEG): container finished" podID="3eebbcd6-55f4-40a5-89b3-ff5ecbddde48" containerID="363465de5b17cab3325aecb1090dcefe6b9a92cb2ff857edf945cd093accae78" exitCode=0 Dec 09 16:08:47 crc kubenswrapper[4894]: I1209 16:08:47.408862 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l" event={"ID":"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48","Type":"ContainerDied","Data":"363465de5b17cab3325aecb1090dcefe6b9a92cb2ff857edf945cd093accae78"} Dec 09 16:08:47 crc kubenswrapper[4894]: I1209 16:08:47.413822 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-64fzw" event={"ID":"ca3465d2-9009-4eb8-ab31-57af05272571","Type":"ContainerStarted","Data":"8fdfaa1a444b39ede5d082acb70b4d73845df66e301acae931e859dd3888b1b5"} Dec 09 16:08:47 crc kubenswrapper[4894]: I1209 16:08:47.460974 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-64fzw" podStartSLOduration=3.264463416 podStartE2EDuration="6.460937964s" podCreationTimestamp="2025-12-09 16:08:41 +0000 UTC" firstStartedPulling="2025-12-09 16:08:43.359986737 +0000 UTC m=+2217.679197426" lastFinishedPulling="2025-12-09 16:08:46.556461305 +0000 UTC m=+2220.875671974" observedRunningTime="2025-12-09 16:08:47.452284709 +0000 UTC m=+2221.771495418" watchObservedRunningTime="2025-12-09 16:08:47.460937964 +0000 UTC m=+2221.780148653" Dec 09 16:08:48 crc kubenswrapper[4894]: I1209 16:08:48.889030 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l" Dec 09 16:08:48 crc kubenswrapper[4894]: I1209 16:08:48.967293 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-ceph\") pod \"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48\" (UID: \"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48\") " Dec 09 16:08:48 crc kubenswrapper[4894]: I1209 16:08:48.967543 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-ssh-key\") pod \"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48\" (UID: \"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48\") " Dec 09 16:08:48 crc kubenswrapper[4894]: I1209 16:08:48.967571 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vvp48\" (UniqueName: \"kubernetes.io/projected/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-kube-api-access-vvp48\") pod \"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48\" (UID: \"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48\") " Dec 09 16:08:48 crc kubenswrapper[4894]: I1209 16:08:48.967611 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-inventory\") pod \"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48\" (UID: \"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48\") " Dec 09 16:08:48 crc kubenswrapper[4894]: I1209 16:08:48.974994 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-ceph" (OuterVolumeSpecName: "ceph") pod "3eebbcd6-55f4-40a5-89b3-ff5ecbddde48" (UID: "3eebbcd6-55f4-40a5-89b3-ff5ecbddde48"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:08:48 crc kubenswrapper[4894]: I1209 16:08:48.975536 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-kube-api-access-vvp48" (OuterVolumeSpecName: "kube-api-access-vvp48") pod "3eebbcd6-55f4-40a5-89b3-ff5ecbddde48" (UID: "3eebbcd6-55f4-40a5-89b3-ff5ecbddde48"). InnerVolumeSpecName "kube-api-access-vvp48". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.013421 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "3eebbcd6-55f4-40a5-89b3-ff5ecbddde48" (UID: "3eebbcd6-55f4-40a5-89b3-ff5ecbddde48"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.024900 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-inventory" (OuterVolumeSpecName: "inventory") pod "3eebbcd6-55f4-40a5-89b3-ff5ecbddde48" (UID: "3eebbcd6-55f4-40a5-89b3-ff5ecbddde48"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.070304 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.070364 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.070381 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.070397 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vvp48\" (UniqueName: \"kubernetes.io/projected/3eebbcd6-55f4-40a5-89b3-ff5ecbddde48-kube-api-access-vvp48\") on node \"crc\" DevicePath \"\"" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.436060 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l" event={"ID":"3eebbcd6-55f4-40a5-89b3-ff5ecbddde48","Type":"ContainerDied","Data":"df6abb6914f00efc776595cbc32ea24464012968d7a9055d96c57223348b2ebf"} Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.436113 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df6abb6914f00efc776595cbc32ea24464012968d7a9055d96c57223348b2ebf" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.436113 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.546525 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg"] Dec 09 16:08:49 crc kubenswrapper[4894]: E1209 16:08:49.547018 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3eebbcd6-55f4-40a5-89b3-ff5ecbddde48" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.547047 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="3eebbcd6-55f4-40a5-89b3-ff5ecbddde48" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.547274 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="3eebbcd6-55f4-40a5-89b3-ff5ecbddde48" containerName="configure-network-edpm-deployment-openstack-edpm-ipam" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.547984 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.550300 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.550587 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.550753 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.550913 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.550965 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.558012 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg"] Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.584921 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-62lbr\" (UniqueName: \"kubernetes.io/projected/ac0dac47-1921-41cf-bc20-c0259c1f75d8-kube-api-access-62lbr\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg\" (UID: \"ac0dac47-1921-41cf-bc20-c0259c1f75d8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.585158 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac0dac47-1921-41cf-bc20-c0259c1f75d8-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg\" (UID: \"ac0dac47-1921-41cf-bc20-c0259c1f75d8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.585195 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ac0dac47-1921-41cf-bc20-c0259c1f75d8-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg\" (UID: \"ac0dac47-1921-41cf-bc20-c0259c1f75d8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.585229 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac0dac47-1921-41cf-bc20-c0259c1f75d8-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg\" (UID: \"ac0dac47-1921-41cf-bc20-c0259c1f75d8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.686453 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac0dac47-1921-41cf-bc20-c0259c1f75d8-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg\" (UID: \"ac0dac47-1921-41cf-bc20-c0259c1f75d8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.686778 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ac0dac47-1921-41cf-bc20-c0259c1f75d8-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg\" (UID: \"ac0dac47-1921-41cf-bc20-c0259c1f75d8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.686802 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac0dac47-1921-41cf-bc20-c0259c1f75d8-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg\" (UID: \"ac0dac47-1921-41cf-bc20-c0259c1f75d8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.686867 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-62lbr\" (UniqueName: \"kubernetes.io/projected/ac0dac47-1921-41cf-bc20-c0259c1f75d8-kube-api-access-62lbr\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg\" (UID: \"ac0dac47-1921-41cf-bc20-c0259c1f75d8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.690378 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac0dac47-1921-41cf-bc20-c0259c1f75d8-inventory\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg\" (UID: \"ac0dac47-1921-41cf-bc20-c0259c1f75d8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.690417 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac0dac47-1921-41cf-bc20-c0259c1f75d8-ssh-key\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg\" (UID: \"ac0dac47-1921-41cf-bc20-c0259c1f75d8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.690862 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ac0dac47-1921-41cf-bc20-c0259c1f75d8-ceph\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg\" (UID: \"ac0dac47-1921-41cf-bc20-c0259c1f75d8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.701451 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-62lbr\" (UniqueName: \"kubernetes.io/projected/ac0dac47-1921-41cf-bc20-c0259c1f75d8-kube-api-access-62lbr\") pod \"validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg\" (UID: \"ac0dac47-1921-41cf-bc20-c0259c1f75d8\") " pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg" Dec 09 16:08:49 crc kubenswrapper[4894]: I1209 16:08:49.890954 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg" Dec 09 16:08:50 crc kubenswrapper[4894]: I1209 16:08:50.447607 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg"] Dec 09 16:08:50 crc kubenswrapper[4894]: W1209 16:08:50.452821 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac0dac47_1921_41cf_bc20_c0259c1f75d8.slice/crio-5fcec18c8203584e94e656aa48c99a570907e27b0268e8711e9cec6bfcd61245 WatchSource:0}: Error finding container 5fcec18c8203584e94e656aa48c99a570907e27b0268e8711e9cec6bfcd61245: Status 404 returned error can't find the container with id 5fcec18c8203584e94e656aa48c99a570907e27b0268e8711e9cec6bfcd61245 Dec 09 16:08:51 crc kubenswrapper[4894]: I1209 16:08:51.460613 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg" event={"ID":"ac0dac47-1921-41cf-bc20-c0259c1f75d8","Type":"ContainerStarted","Data":"5fcec18c8203584e94e656aa48c99a570907e27b0268e8711e9cec6bfcd61245"} Dec 09 16:08:52 crc kubenswrapper[4894]: I1209 16:08:52.147606 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-64fzw" Dec 09 16:08:52 crc kubenswrapper[4894]: I1209 16:08:52.147994 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-64fzw" Dec 09 16:08:52 crc kubenswrapper[4894]: I1209 16:08:52.172187 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cb7w7"] Dec 09 16:08:52 crc kubenswrapper[4894]: I1209 16:08:52.175350 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cb7w7" Dec 09 16:08:52 crc kubenswrapper[4894]: I1209 16:08:52.188857 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cb7w7"] Dec 09 16:08:52 crc kubenswrapper[4894]: I1209 16:08:52.239063 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc640866-0434-4a66-a7e2-fbfb561c290f-utilities\") pod \"community-operators-cb7w7\" (UID: \"cc640866-0434-4a66-a7e2-fbfb561c290f\") " pod="openshift-marketplace/community-operators-cb7w7" Dec 09 16:08:52 crc kubenswrapper[4894]: I1209 16:08:52.239259 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p5l7v\" (UniqueName: \"kubernetes.io/projected/cc640866-0434-4a66-a7e2-fbfb561c290f-kube-api-access-p5l7v\") pod \"community-operators-cb7w7\" (UID: \"cc640866-0434-4a66-a7e2-fbfb561c290f\") " pod="openshift-marketplace/community-operators-cb7w7" Dec 09 16:08:52 crc kubenswrapper[4894]: I1209 16:08:52.239295 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc640866-0434-4a66-a7e2-fbfb561c290f-catalog-content\") pod \"community-operators-cb7w7\" (UID: \"cc640866-0434-4a66-a7e2-fbfb561c290f\") " pod="openshift-marketplace/community-operators-cb7w7" Dec 09 16:08:52 crc kubenswrapper[4894]: I1209 16:08:52.258837 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-64fzw" Dec 09 16:08:52 crc kubenswrapper[4894]: I1209 16:08:52.341330 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p5l7v\" (UniqueName: \"kubernetes.io/projected/cc640866-0434-4a66-a7e2-fbfb561c290f-kube-api-access-p5l7v\") pod \"community-operators-cb7w7\" (UID: \"cc640866-0434-4a66-a7e2-fbfb561c290f\") " pod="openshift-marketplace/community-operators-cb7w7" Dec 09 16:08:52 crc kubenswrapper[4894]: I1209 16:08:52.341410 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc640866-0434-4a66-a7e2-fbfb561c290f-catalog-content\") pod \"community-operators-cb7w7\" (UID: \"cc640866-0434-4a66-a7e2-fbfb561c290f\") " pod="openshift-marketplace/community-operators-cb7w7" Dec 09 16:08:52 crc kubenswrapper[4894]: I1209 16:08:52.341519 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc640866-0434-4a66-a7e2-fbfb561c290f-utilities\") pod \"community-operators-cb7w7\" (UID: \"cc640866-0434-4a66-a7e2-fbfb561c290f\") " pod="openshift-marketplace/community-operators-cb7w7" Dec 09 16:08:52 crc kubenswrapper[4894]: I1209 16:08:52.342200 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc640866-0434-4a66-a7e2-fbfb561c290f-catalog-content\") pod \"community-operators-cb7w7\" (UID: \"cc640866-0434-4a66-a7e2-fbfb561c290f\") " pod="openshift-marketplace/community-operators-cb7w7" Dec 09 16:08:52 crc kubenswrapper[4894]: I1209 16:08:52.342314 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc640866-0434-4a66-a7e2-fbfb561c290f-utilities\") pod \"community-operators-cb7w7\" (UID: \"cc640866-0434-4a66-a7e2-fbfb561c290f\") " pod="openshift-marketplace/community-operators-cb7w7" Dec 09 16:08:52 crc kubenswrapper[4894]: I1209 16:08:52.364838 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p5l7v\" (UniqueName: \"kubernetes.io/projected/cc640866-0434-4a66-a7e2-fbfb561c290f-kube-api-access-p5l7v\") pod \"community-operators-cb7w7\" (UID: \"cc640866-0434-4a66-a7e2-fbfb561c290f\") " pod="openshift-marketplace/community-operators-cb7w7" Dec 09 16:08:52 crc kubenswrapper[4894]: I1209 16:08:52.470319 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg" event={"ID":"ac0dac47-1921-41cf-bc20-c0259c1f75d8","Type":"ContainerStarted","Data":"9427b4f3b80aaca535268fe47017e5eaf0242d49558328938d373190d68179e7"} Dec 09 16:08:52 crc kubenswrapper[4894]: I1209 16:08:52.497815 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg" podStartSLOduration=2.271363035 podStartE2EDuration="3.497794866s" podCreationTimestamp="2025-12-09 16:08:49 +0000 UTC" firstStartedPulling="2025-12-09 16:08:50.457078397 +0000 UTC m=+2224.776289066" lastFinishedPulling="2025-12-09 16:08:51.683510178 +0000 UTC m=+2226.002720897" observedRunningTime="2025-12-09 16:08:52.493067947 +0000 UTC m=+2226.812278636" watchObservedRunningTime="2025-12-09 16:08:52.497794866 +0000 UTC m=+2226.817005535" Dec 09 16:08:52 crc kubenswrapper[4894]: I1209 16:08:52.512400 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cb7w7" Dec 09 16:08:52 crc kubenswrapper[4894]: I1209 16:08:52.549081 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-64fzw" Dec 09 16:08:53 crc kubenswrapper[4894]: I1209 16:08:53.018036 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cb7w7"] Dec 09 16:08:53 crc kubenswrapper[4894]: I1209 16:08:53.481164 4894 generic.go:334] "Generic (PLEG): container finished" podID="cc640866-0434-4a66-a7e2-fbfb561c290f" containerID="a9070d7822e20ecf6592ccaf08869fc50eaffb0418b09b1e4e05a64bc4e9ab68" exitCode=0 Dec 09 16:08:53 crc kubenswrapper[4894]: I1209 16:08:53.481228 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cb7w7" event={"ID":"cc640866-0434-4a66-a7e2-fbfb561c290f","Type":"ContainerDied","Data":"a9070d7822e20ecf6592ccaf08869fc50eaffb0418b09b1e4e05a64bc4e9ab68"} Dec 09 16:08:53 crc kubenswrapper[4894]: I1209 16:08:53.481298 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cb7w7" event={"ID":"cc640866-0434-4a66-a7e2-fbfb561c290f","Type":"ContainerStarted","Data":"d4703a96b712628aa7487a9134c557fd8efe9d03f9d1d71613d560f6f8084f58"} Dec 09 16:08:54 crc kubenswrapper[4894]: I1209 16:08:54.536617 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-64fzw"] Dec 09 16:08:54 crc kubenswrapper[4894]: I1209 16:08:54.537281 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-64fzw" podUID="ca3465d2-9009-4eb8-ab31-57af05272571" containerName="registry-server" containerID="cri-o://8fdfaa1a444b39ede5d082acb70b4d73845df66e301acae931e859dd3888b1b5" gracePeriod=2 Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.010499 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-64fzw" Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.089415 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca3465d2-9009-4eb8-ab31-57af05272571-utilities\") pod \"ca3465d2-9009-4eb8-ab31-57af05272571\" (UID: \"ca3465d2-9009-4eb8-ab31-57af05272571\") " Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.089505 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca3465d2-9009-4eb8-ab31-57af05272571-catalog-content\") pod \"ca3465d2-9009-4eb8-ab31-57af05272571\" (UID: \"ca3465d2-9009-4eb8-ab31-57af05272571\") " Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.089592 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4sd7\" (UniqueName: \"kubernetes.io/projected/ca3465d2-9009-4eb8-ab31-57af05272571-kube-api-access-w4sd7\") pod \"ca3465d2-9009-4eb8-ab31-57af05272571\" (UID: \"ca3465d2-9009-4eb8-ab31-57af05272571\") " Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.090919 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca3465d2-9009-4eb8-ab31-57af05272571-utilities" (OuterVolumeSpecName: "utilities") pod "ca3465d2-9009-4eb8-ab31-57af05272571" (UID: "ca3465d2-9009-4eb8-ab31-57af05272571"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.096076 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ca3465d2-9009-4eb8-ab31-57af05272571-kube-api-access-w4sd7" (OuterVolumeSpecName: "kube-api-access-w4sd7") pod "ca3465d2-9009-4eb8-ab31-57af05272571" (UID: "ca3465d2-9009-4eb8-ab31-57af05272571"). InnerVolumeSpecName "kube-api-access-w4sd7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.110985 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ca3465d2-9009-4eb8-ab31-57af05272571-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "ca3465d2-9009-4eb8-ab31-57af05272571" (UID: "ca3465d2-9009-4eb8-ab31-57af05272571"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.192269 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4sd7\" (UniqueName: \"kubernetes.io/projected/ca3465d2-9009-4eb8-ab31-57af05272571-kube-api-access-w4sd7\") on node \"crc\" DevicePath \"\"" Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.192542 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ca3465d2-9009-4eb8-ab31-57af05272571-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.192552 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ca3465d2-9009-4eb8-ab31-57af05272571-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.511156 4894 generic.go:334] "Generic (PLEG): container finished" podID="ca3465d2-9009-4eb8-ab31-57af05272571" containerID="8fdfaa1a444b39ede5d082acb70b4d73845df66e301acae931e859dd3888b1b5" exitCode=0 Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.511251 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-64fzw" event={"ID":"ca3465d2-9009-4eb8-ab31-57af05272571","Type":"ContainerDied","Data":"8fdfaa1a444b39ede5d082acb70b4d73845df66e301acae931e859dd3888b1b5"} Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.511429 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-64fzw" event={"ID":"ca3465d2-9009-4eb8-ab31-57af05272571","Type":"ContainerDied","Data":"5b4f7eeb0cc1ede9dfbcbb5159ebdedcbadf7ef5deb884e97a339237b363f11e"} Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.511460 4894 scope.go:117] "RemoveContainer" containerID="8fdfaa1a444b39ede5d082acb70b4d73845df66e301acae931e859dd3888b1b5" Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.511286 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-64fzw" Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.516855 4894 generic.go:334] "Generic (PLEG): container finished" podID="cc640866-0434-4a66-a7e2-fbfb561c290f" containerID="c881e80b87de9234f90cde669f1c04837076e50ee7445532e746497d67cb83cc" exitCode=0 Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.516895 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cb7w7" event={"ID":"cc640866-0434-4a66-a7e2-fbfb561c290f","Type":"ContainerDied","Data":"c881e80b87de9234f90cde669f1c04837076e50ee7445532e746497d67cb83cc"} Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.559872 4894 scope.go:117] "RemoveContainer" containerID="3cd6d564db6a03c6f24683b35bb3ca0930d2160a0a86c516ce9862494ba3b1bf" Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.588225 4894 scope.go:117] "RemoveContainer" containerID="6df924c4219dbb8957d31f4b29c815434b186b9ea684eb94b04c7c852505a104" Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.590870 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-64fzw"] Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.603382 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-64fzw"] Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.611842 4894 scope.go:117] "RemoveContainer" containerID="8fdfaa1a444b39ede5d082acb70b4d73845df66e301acae931e859dd3888b1b5" Dec 09 16:08:55 crc kubenswrapper[4894]: E1209 16:08:55.612480 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8fdfaa1a444b39ede5d082acb70b4d73845df66e301acae931e859dd3888b1b5\": container with ID starting with 8fdfaa1a444b39ede5d082acb70b4d73845df66e301acae931e859dd3888b1b5 not found: ID does not exist" containerID="8fdfaa1a444b39ede5d082acb70b4d73845df66e301acae931e859dd3888b1b5" Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.612568 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8fdfaa1a444b39ede5d082acb70b4d73845df66e301acae931e859dd3888b1b5"} err="failed to get container status \"8fdfaa1a444b39ede5d082acb70b4d73845df66e301acae931e859dd3888b1b5\": rpc error: code = NotFound desc = could not find container \"8fdfaa1a444b39ede5d082acb70b4d73845df66e301acae931e859dd3888b1b5\": container with ID starting with 8fdfaa1a444b39ede5d082acb70b4d73845df66e301acae931e859dd3888b1b5 not found: ID does not exist" Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.612610 4894 scope.go:117] "RemoveContainer" containerID="3cd6d564db6a03c6f24683b35bb3ca0930d2160a0a86c516ce9862494ba3b1bf" Dec 09 16:08:55 crc kubenswrapper[4894]: E1209 16:08:55.613301 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3cd6d564db6a03c6f24683b35bb3ca0930d2160a0a86c516ce9862494ba3b1bf\": container with ID starting with 3cd6d564db6a03c6f24683b35bb3ca0930d2160a0a86c516ce9862494ba3b1bf not found: ID does not exist" containerID="3cd6d564db6a03c6f24683b35bb3ca0930d2160a0a86c516ce9862494ba3b1bf" Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.613440 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3cd6d564db6a03c6f24683b35bb3ca0930d2160a0a86c516ce9862494ba3b1bf"} err="failed to get container status \"3cd6d564db6a03c6f24683b35bb3ca0930d2160a0a86c516ce9862494ba3b1bf\": rpc error: code = NotFound desc = could not find container \"3cd6d564db6a03c6f24683b35bb3ca0930d2160a0a86c516ce9862494ba3b1bf\": container with ID starting with 3cd6d564db6a03c6f24683b35bb3ca0930d2160a0a86c516ce9862494ba3b1bf not found: ID does not exist" Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.613577 4894 scope.go:117] "RemoveContainer" containerID="6df924c4219dbb8957d31f4b29c815434b186b9ea684eb94b04c7c852505a104" Dec 09 16:08:55 crc kubenswrapper[4894]: E1209 16:08:55.614034 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6df924c4219dbb8957d31f4b29c815434b186b9ea684eb94b04c7c852505a104\": container with ID starting with 6df924c4219dbb8957d31f4b29c815434b186b9ea684eb94b04c7c852505a104 not found: ID does not exist" containerID="6df924c4219dbb8957d31f4b29c815434b186b9ea684eb94b04c7c852505a104" Dec 09 16:08:55 crc kubenswrapper[4894]: I1209 16:08:55.614154 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6df924c4219dbb8957d31f4b29c815434b186b9ea684eb94b04c7c852505a104"} err="failed to get container status \"6df924c4219dbb8957d31f4b29c815434b186b9ea684eb94b04c7c852505a104\": rpc error: code = NotFound desc = could not find container \"6df924c4219dbb8957d31f4b29c815434b186b9ea684eb94b04c7c852505a104\": container with ID starting with 6df924c4219dbb8957d31f4b29c815434b186b9ea684eb94b04c7c852505a104 not found: ID does not exist" Dec 09 16:08:56 crc kubenswrapper[4894]: I1209 16:08:56.121966 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ca3465d2-9009-4eb8-ab31-57af05272571" path="/var/lib/kubelet/pods/ca3465d2-9009-4eb8-ab31-57af05272571/volumes" Dec 09 16:08:57 crc kubenswrapper[4894]: I1209 16:08:57.544856 4894 generic.go:334] "Generic (PLEG): container finished" podID="ac0dac47-1921-41cf-bc20-c0259c1f75d8" containerID="9427b4f3b80aaca535268fe47017e5eaf0242d49558328938d373190d68179e7" exitCode=0 Dec 09 16:08:57 crc kubenswrapper[4894]: I1209 16:08:57.544969 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg" event={"ID":"ac0dac47-1921-41cf-bc20-c0259c1f75d8","Type":"ContainerDied","Data":"9427b4f3b80aaca535268fe47017e5eaf0242d49558328938d373190d68179e7"} Dec 09 16:08:57 crc kubenswrapper[4894]: I1209 16:08:57.551402 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cb7w7" event={"ID":"cc640866-0434-4a66-a7e2-fbfb561c290f","Type":"ContainerStarted","Data":"d8bd59163bdd807cda90f90bb5b0e253e81d0f4c7990fe4e60c6232424cfe20b"} Dec 09 16:08:57 crc kubenswrapper[4894]: I1209 16:08:57.599111 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cb7w7" podStartSLOduration=2.530117235 podStartE2EDuration="5.599092148s" podCreationTimestamp="2025-12-09 16:08:52 +0000 UTC" firstStartedPulling="2025-12-09 16:08:53.483014939 +0000 UTC m=+2227.802225608" lastFinishedPulling="2025-12-09 16:08:56.551989852 +0000 UTC m=+2230.871200521" observedRunningTime="2025-12-09 16:08:57.595179862 +0000 UTC m=+2231.914390561" watchObservedRunningTime="2025-12-09 16:08:57.599092148 +0000 UTC m=+2231.918302837" Dec 09 16:08:58 crc kubenswrapper[4894]: I1209 16:08:58.108507 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:08:58 crc kubenswrapper[4894]: E1209 16:08:58.109149 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.014214 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.072915 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac0dac47-1921-41cf-bc20-c0259c1f75d8-inventory\") pod \"ac0dac47-1921-41cf-bc20-c0259c1f75d8\" (UID: \"ac0dac47-1921-41cf-bc20-c0259c1f75d8\") " Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.073197 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-62lbr\" (UniqueName: \"kubernetes.io/projected/ac0dac47-1921-41cf-bc20-c0259c1f75d8-kube-api-access-62lbr\") pod \"ac0dac47-1921-41cf-bc20-c0259c1f75d8\" (UID: \"ac0dac47-1921-41cf-bc20-c0259c1f75d8\") " Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.073228 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac0dac47-1921-41cf-bc20-c0259c1f75d8-ssh-key\") pod \"ac0dac47-1921-41cf-bc20-c0259c1f75d8\" (UID: \"ac0dac47-1921-41cf-bc20-c0259c1f75d8\") " Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.073286 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ac0dac47-1921-41cf-bc20-c0259c1f75d8-ceph\") pod \"ac0dac47-1921-41cf-bc20-c0259c1f75d8\" (UID: \"ac0dac47-1921-41cf-bc20-c0259c1f75d8\") " Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.082214 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ac0dac47-1921-41cf-bc20-c0259c1f75d8-kube-api-access-62lbr" (OuterVolumeSpecName: "kube-api-access-62lbr") pod "ac0dac47-1921-41cf-bc20-c0259c1f75d8" (UID: "ac0dac47-1921-41cf-bc20-c0259c1f75d8"). InnerVolumeSpecName "kube-api-access-62lbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.082402 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac0dac47-1921-41cf-bc20-c0259c1f75d8-ceph" (OuterVolumeSpecName: "ceph") pod "ac0dac47-1921-41cf-bc20-c0259c1f75d8" (UID: "ac0dac47-1921-41cf-bc20-c0259c1f75d8"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.106150 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac0dac47-1921-41cf-bc20-c0259c1f75d8-inventory" (OuterVolumeSpecName: "inventory") pod "ac0dac47-1921-41cf-bc20-c0259c1f75d8" (UID: "ac0dac47-1921-41cf-bc20-c0259c1f75d8"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.122853 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ac0dac47-1921-41cf-bc20-c0259c1f75d8-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "ac0dac47-1921-41cf-bc20-c0259c1f75d8" (UID: "ac0dac47-1921-41cf-bc20-c0259c1f75d8"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.176234 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-62lbr\" (UniqueName: \"kubernetes.io/projected/ac0dac47-1921-41cf-bc20-c0259c1f75d8-kube-api-access-62lbr\") on node \"crc\" DevicePath \"\"" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.177423 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/ac0dac47-1921-41cf-bc20-c0259c1f75d8-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.177521 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/ac0dac47-1921-41cf-bc20-c0259c1f75d8-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.177598 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/ac0dac47-1921-41cf-bc20-c0259c1f75d8-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.569851 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg" event={"ID":"ac0dac47-1921-41cf-bc20-c0259c1f75d8","Type":"ContainerDied","Data":"5fcec18c8203584e94e656aa48c99a570907e27b0268e8711e9cec6bfcd61245"} Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.569892 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.569896 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5fcec18c8203584e94e656aa48c99a570907e27b0268e8711e9cec6bfcd61245" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.668459 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5"] Dec 09 16:08:59 crc kubenswrapper[4894]: E1209 16:08:59.669450 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca3465d2-9009-4eb8-ab31-57af05272571" containerName="extract-content" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.669474 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca3465d2-9009-4eb8-ab31-57af05272571" containerName="extract-content" Dec 09 16:08:59 crc kubenswrapper[4894]: E1209 16:08:59.669543 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca3465d2-9009-4eb8-ab31-57af05272571" containerName="extract-utilities" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.669565 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca3465d2-9009-4eb8-ab31-57af05272571" containerName="extract-utilities" Dec 09 16:08:59 crc kubenswrapper[4894]: E1209 16:08:59.669617 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ca3465d2-9009-4eb8-ab31-57af05272571" containerName="registry-server" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.669653 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ca3465d2-9009-4eb8-ab31-57af05272571" containerName="registry-server" Dec 09 16:08:59 crc kubenswrapper[4894]: E1209 16:08:59.669695 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ac0dac47-1921-41cf-bc20-c0259c1f75d8" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.669706 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ac0dac47-1921-41cf-bc20-c0259c1f75d8" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.670117 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="ca3465d2-9009-4eb8-ab31-57af05272571" containerName="registry-server" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.670170 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="ac0dac47-1921-41cf-bc20-c0259c1f75d8" containerName="validate-network-edpm-deployment-openstack-edpm-ipam" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.671429 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.676346 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.676528 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.676574 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.676630 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.676804 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.682732 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5"] Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.685171 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/031a7ee5-ae47-45d2-9fd2-7b8beef85242-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vg8f5\" (UID: \"031a7ee5-ae47-45d2-9fd2-7b8beef85242\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.685290 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gh2h2\" (UniqueName: \"kubernetes.io/projected/031a7ee5-ae47-45d2-9fd2-7b8beef85242-kube-api-access-gh2h2\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vg8f5\" (UID: \"031a7ee5-ae47-45d2-9fd2-7b8beef85242\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.685324 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/031a7ee5-ae47-45d2-9fd2-7b8beef85242-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vg8f5\" (UID: \"031a7ee5-ae47-45d2-9fd2-7b8beef85242\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.685400 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/031a7ee5-ae47-45d2-9fd2-7b8beef85242-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vg8f5\" (UID: \"031a7ee5-ae47-45d2-9fd2-7b8beef85242\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.786552 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/031a7ee5-ae47-45d2-9fd2-7b8beef85242-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vg8f5\" (UID: \"031a7ee5-ae47-45d2-9fd2-7b8beef85242\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.786598 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/031a7ee5-ae47-45d2-9fd2-7b8beef85242-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vg8f5\" (UID: \"031a7ee5-ae47-45d2-9fd2-7b8beef85242\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.786699 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gh2h2\" (UniqueName: \"kubernetes.io/projected/031a7ee5-ae47-45d2-9fd2-7b8beef85242-kube-api-access-gh2h2\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vg8f5\" (UID: \"031a7ee5-ae47-45d2-9fd2-7b8beef85242\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.786725 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/031a7ee5-ae47-45d2-9fd2-7b8beef85242-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vg8f5\" (UID: \"031a7ee5-ae47-45d2-9fd2-7b8beef85242\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.790163 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/031a7ee5-ae47-45d2-9fd2-7b8beef85242-ssh-key\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vg8f5\" (UID: \"031a7ee5-ae47-45d2-9fd2-7b8beef85242\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.791278 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/031a7ee5-ae47-45d2-9fd2-7b8beef85242-inventory\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vg8f5\" (UID: \"031a7ee5-ae47-45d2-9fd2-7b8beef85242\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.793501 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/031a7ee5-ae47-45d2-9fd2-7b8beef85242-ceph\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vg8f5\" (UID: \"031a7ee5-ae47-45d2-9fd2-7b8beef85242\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.803317 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gh2h2\" (UniqueName: \"kubernetes.io/projected/031a7ee5-ae47-45d2-9fd2-7b8beef85242-kube-api-access-gh2h2\") pod \"install-os-edpm-deployment-openstack-edpm-ipam-vg8f5\" (UID: \"031a7ee5-ae47-45d2-9fd2-7b8beef85242\") " pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5" Dec 09 16:08:59 crc kubenswrapper[4894]: I1209 16:08:59.993344 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5" Dec 09 16:09:00 crc kubenswrapper[4894]: I1209 16:09:00.538488 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5"] Dec 09 16:09:00 crc kubenswrapper[4894]: W1209 16:09:00.540523 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod031a7ee5_ae47_45d2_9fd2_7b8beef85242.slice/crio-ebe729a42635b1c94cf72a8c81d5792578dce268d7824d0f68b004c1c5c303b7 WatchSource:0}: Error finding container ebe729a42635b1c94cf72a8c81d5792578dce268d7824d0f68b004c1c5c303b7: Status 404 returned error can't find the container with id ebe729a42635b1c94cf72a8c81d5792578dce268d7824d0f68b004c1c5c303b7 Dec 09 16:09:00 crc kubenswrapper[4894]: I1209 16:09:00.579323 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5" event={"ID":"031a7ee5-ae47-45d2-9fd2-7b8beef85242","Type":"ContainerStarted","Data":"ebe729a42635b1c94cf72a8c81d5792578dce268d7824d0f68b004c1c5c303b7"} Dec 09 16:09:02 crc kubenswrapper[4894]: I1209 16:09:02.513013 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cb7w7" Dec 09 16:09:02 crc kubenswrapper[4894]: I1209 16:09:02.513589 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cb7w7" Dec 09 16:09:02 crc kubenswrapper[4894]: I1209 16:09:02.567790 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cb7w7" Dec 09 16:09:02 crc kubenswrapper[4894]: I1209 16:09:02.603088 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5" event={"ID":"031a7ee5-ae47-45d2-9fd2-7b8beef85242","Type":"ContainerStarted","Data":"9b0cb41955214bbb6747956045816e7700c5548a71079dd8ec002bad649b87c5"} Dec 09 16:09:02 crc kubenswrapper[4894]: I1209 16:09:02.626927 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5" podStartSLOduration=2.408298045 podStartE2EDuration="3.626906044s" podCreationTimestamp="2025-12-09 16:08:59 +0000 UTC" firstStartedPulling="2025-12-09 16:09:00.543256748 +0000 UTC m=+2234.862467427" lastFinishedPulling="2025-12-09 16:09:01.761864697 +0000 UTC m=+2236.081075426" observedRunningTime="2025-12-09 16:09:02.620342945 +0000 UTC m=+2236.939553614" watchObservedRunningTime="2025-12-09 16:09:02.626906044 +0000 UTC m=+2236.946116713" Dec 09 16:09:02 crc kubenswrapper[4894]: I1209 16:09:02.652660 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cb7w7" Dec 09 16:09:02 crc kubenswrapper[4894]: I1209 16:09:02.803969 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cb7w7"] Dec 09 16:09:04 crc kubenswrapper[4894]: I1209 16:09:04.628438 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cb7w7" podUID="cc640866-0434-4a66-a7e2-fbfb561c290f" containerName="registry-server" containerID="cri-o://d8bd59163bdd807cda90f90bb5b0e253e81d0f4c7990fe4e60c6232424cfe20b" gracePeriod=2 Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.094423 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cb7w7" Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.205626 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc640866-0434-4a66-a7e2-fbfb561c290f-catalog-content\") pod \"cc640866-0434-4a66-a7e2-fbfb561c290f\" (UID: \"cc640866-0434-4a66-a7e2-fbfb561c290f\") " Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.205752 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc640866-0434-4a66-a7e2-fbfb561c290f-utilities\") pod \"cc640866-0434-4a66-a7e2-fbfb561c290f\" (UID: \"cc640866-0434-4a66-a7e2-fbfb561c290f\") " Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.205783 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p5l7v\" (UniqueName: \"kubernetes.io/projected/cc640866-0434-4a66-a7e2-fbfb561c290f-kube-api-access-p5l7v\") pod \"cc640866-0434-4a66-a7e2-fbfb561c290f\" (UID: \"cc640866-0434-4a66-a7e2-fbfb561c290f\") " Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.207302 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc640866-0434-4a66-a7e2-fbfb561c290f-utilities" (OuterVolumeSpecName: "utilities") pod "cc640866-0434-4a66-a7e2-fbfb561c290f" (UID: "cc640866-0434-4a66-a7e2-fbfb561c290f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.211618 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cc640866-0434-4a66-a7e2-fbfb561c290f-kube-api-access-p5l7v" (OuterVolumeSpecName: "kube-api-access-p5l7v") pod "cc640866-0434-4a66-a7e2-fbfb561c290f" (UID: "cc640866-0434-4a66-a7e2-fbfb561c290f"). InnerVolumeSpecName "kube-api-access-p5l7v". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.267339 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cc640866-0434-4a66-a7e2-fbfb561c290f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "cc640866-0434-4a66-a7e2-fbfb561c290f" (UID: "cc640866-0434-4a66-a7e2-fbfb561c290f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.308656 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cc640866-0434-4a66-a7e2-fbfb561c290f-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.308684 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p5l7v\" (UniqueName: \"kubernetes.io/projected/cc640866-0434-4a66-a7e2-fbfb561c290f-kube-api-access-p5l7v\") on node \"crc\" DevicePath \"\"" Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.308697 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cc640866-0434-4a66-a7e2-fbfb561c290f-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.641847 4894 generic.go:334] "Generic (PLEG): container finished" podID="cc640866-0434-4a66-a7e2-fbfb561c290f" containerID="d8bd59163bdd807cda90f90bb5b0e253e81d0f4c7990fe4e60c6232424cfe20b" exitCode=0 Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.641941 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cb7w7" event={"ID":"cc640866-0434-4a66-a7e2-fbfb561c290f","Type":"ContainerDied","Data":"d8bd59163bdd807cda90f90bb5b0e253e81d0f4c7990fe4e60c6232424cfe20b"} Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.641966 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cb7w7" Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.643171 4894 scope.go:117] "RemoveContainer" containerID="d8bd59163bdd807cda90f90bb5b0e253e81d0f4c7990fe4e60c6232424cfe20b" Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.643077 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cb7w7" event={"ID":"cc640866-0434-4a66-a7e2-fbfb561c290f","Type":"ContainerDied","Data":"d4703a96b712628aa7487a9134c557fd8efe9d03f9d1d71613d560f6f8084f58"} Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.674821 4894 scope.go:117] "RemoveContainer" containerID="c881e80b87de9234f90cde669f1c04837076e50ee7445532e746497d67cb83cc" Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.703198 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cb7w7"] Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.713893 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cb7w7"] Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.736802 4894 scope.go:117] "RemoveContainer" containerID="a9070d7822e20ecf6592ccaf08869fc50eaffb0418b09b1e4e05a64bc4e9ab68" Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.758328 4894 scope.go:117] "RemoveContainer" containerID="d8bd59163bdd807cda90f90bb5b0e253e81d0f4c7990fe4e60c6232424cfe20b" Dec 09 16:09:05 crc kubenswrapper[4894]: E1209 16:09:05.758772 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8bd59163bdd807cda90f90bb5b0e253e81d0f4c7990fe4e60c6232424cfe20b\": container with ID starting with d8bd59163bdd807cda90f90bb5b0e253e81d0f4c7990fe4e60c6232424cfe20b not found: ID does not exist" containerID="d8bd59163bdd807cda90f90bb5b0e253e81d0f4c7990fe4e60c6232424cfe20b" Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.758806 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8bd59163bdd807cda90f90bb5b0e253e81d0f4c7990fe4e60c6232424cfe20b"} err="failed to get container status \"d8bd59163bdd807cda90f90bb5b0e253e81d0f4c7990fe4e60c6232424cfe20b\": rpc error: code = NotFound desc = could not find container \"d8bd59163bdd807cda90f90bb5b0e253e81d0f4c7990fe4e60c6232424cfe20b\": container with ID starting with d8bd59163bdd807cda90f90bb5b0e253e81d0f4c7990fe4e60c6232424cfe20b not found: ID does not exist" Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.758827 4894 scope.go:117] "RemoveContainer" containerID="c881e80b87de9234f90cde669f1c04837076e50ee7445532e746497d67cb83cc" Dec 09 16:09:05 crc kubenswrapper[4894]: E1209 16:09:05.759231 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c881e80b87de9234f90cde669f1c04837076e50ee7445532e746497d67cb83cc\": container with ID starting with c881e80b87de9234f90cde669f1c04837076e50ee7445532e746497d67cb83cc not found: ID does not exist" containerID="c881e80b87de9234f90cde669f1c04837076e50ee7445532e746497d67cb83cc" Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.759254 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c881e80b87de9234f90cde669f1c04837076e50ee7445532e746497d67cb83cc"} err="failed to get container status \"c881e80b87de9234f90cde669f1c04837076e50ee7445532e746497d67cb83cc\": rpc error: code = NotFound desc = could not find container \"c881e80b87de9234f90cde669f1c04837076e50ee7445532e746497d67cb83cc\": container with ID starting with c881e80b87de9234f90cde669f1c04837076e50ee7445532e746497d67cb83cc not found: ID does not exist" Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.759269 4894 scope.go:117] "RemoveContainer" containerID="a9070d7822e20ecf6592ccaf08869fc50eaffb0418b09b1e4e05a64bc4e9ab68" Dec 09 16:09:05 crc kubenswrapper[4894]: E1209 16:09:05.759468 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a9070d7822e20ecf6592ccaf08869fc50eaffb0418b09b1e4e05a64bc4e9ab68\": container with ID starting with a9070d7822e20ecf6592ccaf08869fc50eaffb0418b09b1e4e05a64bc4e9ab68 not found: ID does not exist" containerID="a9070d7822e20ecf6592ccaf08869fc50eaffb0418b09b1e4e05a64bc4e9ab68" Dec 09 16:09:05 crc kubenswrapper[4894]: I1209 16:09:05.759489 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a9070d7822e20ecf6592ccaf08869fc50eaffb0418b09b1e4e05a64bc4e9ab68"} err="failed to get container status \"a9070d7822e20ecf6592ccaf08869fc50eaffb0418b09b1e4e05a64bc4e9ab68\": rpc error: code = NotFound desc = could not find container \"a9070d7822e20ecf6592ccaf08869fc50eaffb0418b09b1e4e05a64bc4e9ab68\": container with ID starting with a9070d7822e20ecf6592ccaf08869fc50eaffb0418b09b1e4e05a64bc4e9ab68 not found: ID does not exist" Dec 09 16:09:06 crc kubenswrapper[4894]: I1209 16:09:06.124947 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cc640866-0434-4a66-a7e2-fbfb561c290f" path="/var/lib/kubelet/pods/cc640866-0434-4a66-a7e2-fbfb561c290f/volumes" Dec 09 16:09:12 crc kubenswrapper[4894]: I1209 16:09:12.106844 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:09:12 crc kubenswrapper[4894]: E1209 16:09:12.107680 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:09:26 crc kubenswrapper[4894]: I1209 16:09:26.121506 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:09:26 crc kubenswrapper[4894]: E1209 16:09:26.122790 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:09:38 crc kubenswrapper[4894]: I1209 16:09:38.960989 4894 generic.go:334] "Generic (PLEG): container finished" podID="031a7ee5-ae47-45d2-9fd2-7b8beef85242" containerID="9b0cb41955214bbb6747956045816e7700c5548a71079dd8ec002bad649b87c5" exitCode=0 Dec 09 16:09:38 crc kubenswrapper[4894]: I1209 16:09:38.961081 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5" event={"ID":"031a7ee5-ae47-45d2-9fd2-7b8beef85242","Type":"ContainerDied","Data":"9b0cb41955214bbb6747956045816e7700c5548a71079dd8ec002bad649b87c5"} Dec 09 16:09:40 crc kubenswrapper[4894]: I1209 16:09:40.107088 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:09:40 crc kubenswrapper[4894]: E1209 16:09:40.107662 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:09:40 crc kubenswrapper[4894]: I1209 16:09:40.464845 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5" Dec 09 16:09:40 crc kubenswrapper[4894]: I1209 16:09:40.583584 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/031a7ee5-ae47-45d2-9fd2-7b8beef85242-ceph\") pod \"031a7ee5-ae47-45d2-9fd2-7b8beef85242\" (UID: \"031a7ee5-ae47-45d2-9fd2-7b8beef85242\") " Dec 09 16:09:40 crc kubenswrapper[4894]: I1209 16:09:40.584802 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/031a7ee5-ae47-45d2-9fd2-7b8beef85242-inventory\") pod \"031a7ee5-ae47-45d2-9fd2-7b8beef85242\" (UID: \"031a7ee5-ae47-45d2-9fd2-7b8beef85242\") " Dec 09 16:09:40 crc kubenswrapper[4894]: I1209 16:09:40.584834 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/031a7ee5-ae47-45d2-9fd2-7b8beef85242-ssh-key\") pod \"031a7ee5-ae47-45d2-9fd2-7b8beef85242\" (UID: \"031a7ee5-ae47-45d2-9fd2-7b8beef85242\") " Dec 09 16:09:40 crc kubenswrapper[4894]: I1209 16:09:40.584906 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gh2h2\" (UniqueName: \"kubernetes.io/projected/031a7ee5-ae47-45d2-9fd2-7b8beef85242-kube-api-access-gh2h2\") pod \"031a7ee5-ae47-45d2-9fd2-7b8beef85242\" (UID: \"031a7ee5-ae47-45d2-9fd2-7b8beef85242\") " Dec 09 16:09:40 crc kubenswrapper[4894]: I1209 16:09:40.589938 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/031a7ee5-ae47-45d2-9fd2-7b8beef85242-ceph" (OuterVolumeSpecName: "ceph") pod "031a7ee5-ae47-45d2-9fd2-7b8beef85242" (UID: "031a7ee5-ae47-45d2-9fd2-7b8beef85242"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:09:40 crc kubenswrapper[4894]: I1209 16:09:40.591813 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/031a7ee5-ae47-45d2-9fd2-7b8beef85242-kube-api-access-gh2h2" (OuterVolumeSpecName: "kube-api-access-gh2h2") pod "031a7ee5-ae47-45d2-9fd2-7b8beef85242" (UID: "031a7ee5-ae47-45d2-9fd2-7b8beef85242"). InnerVolumeSpecName "kube-api-access-gh2h2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:09:40 crc kubenswrapper[4894]: I1209 16:09:40.615811 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/031a7ee5-ae47-45d2-9fd2-7b8beef85242-inventory" (OuterVolumeSpecName: "inventory") pod "031a7ee5-ae47-45d2-9fd2-7b8beef85242" (UID: "031a7ee5-ae47-45d2-9fd2-7b8beef85242"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:09:40 crc kubenswrapper[4894]: I1209 16:09:40.619561 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/031a7ee5-ae47-45d2-9fd2-7b8beef85242-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "031a7ee5-ae47-45d2-9fd2-7b8beef85242" (UID: "031a7ee5-ae47-45d2-9fd2-7b8beef85242"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:09:40 crc kubenswrapper[4894]: I1209 16:09:40.686428 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/031a7ee5-ae47-45d2-9fd2-7b8beef85242-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:09:40 crc kubenswrapper[4894]: I1209 16:09:40.686461 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/031a7ee5-ae47-45d2-9fd2-7b8beef85242-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 16:09:40 crc kubenswrapper[4894]: I1209 16:09:40.686471 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/031a7ee5-ae47-45d2-9fd2-7b8beef85242-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:09:40 crc kubenswrapper[4894]: I1209 16:09:40.686483 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gh2h2\" (UniqueName: \"kubernetes.io/projected/031a7ee5-ae47-45d2-9fd2-7b8beef85242-kube-api-access-gh2h2\") on node \"crc\" DevicePath \"\"" Dec 09 16:09:40 crc kubenswrapper[4894]: I1209 16:09:40.981300 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5" event={"ID":"031a7ee5-ae47-45d2-9fd2-7b8beef85242","Type":"ContainerDied","Data":"ebe729a42635b1c94cf72a8c81d5792578dce268d7824d0f68b004c1c5c303b7"} Dec 09 16:09:40 crc kubenswrapper[4894]: I1209 16:09:40.981344 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ebe729a42635b1c94cf72a8c81d5792578dce268d7824d0f68b004c1c5c303b7" Dec 09 16:09:40 crc kubenswrapper[4894]: I1209 16:09:40.981368 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-os-edpm-deployment-openstack-edpm-ipam-vg8f5" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.074623 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn"] Dec 09 16:09:41 crc kubenswrapper[4894]: E1209 16:09:41.075060 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc640866-0434-4a66-a7e2-fbfb561c290f" containerName="registry-server" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.075072 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc640866-0434-4a66-a7e2-fbfb561c290f" containerName="registry-server" Dec 09 16:09:41 crc kubenswrapper[4894]: E1209 16:09:41.075085 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="031a7ee5-ae47-45d2-9fd2-7b8beef85242" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.075092 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="031a7ee5-ae47-45d2-9fd2-7b8beef85242" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:09:41 crc kubenswrapper[4894]: E1209 16:09:41.075105 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc640866-0434-4a66-a7e2-fbfb561c290f" containerName="extract-content" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.075111 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc640866-0434-4a66-a7e2-fbfb561c290f" containerName="extract-content" Dec 09 16:09:41 crc kubenswrapper[4894]: E1209 16:09:41.075125 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cc640866-0434-4a66-a7e2-fbfb561c290f" containerName="extract-utilities" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.075130 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="cc640866-0434-4a66-a7e2-fbfb561c290f" containerName="extract-utilities" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.075276 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="cc640866-0434-4a66-a7e2-fbfb561c290f" containerName="registry-server" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.075301 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="031a7ee5-ae47-45d2-9fd2-7b8beef85242" containerName="install-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.075847 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.078474 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.078755 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.078764 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.080606 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.080836 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.104835 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn"] Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.194888 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn\" (UID: \"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.195608 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ln2hs\" (UniqueName: \"kubernetes.io/projected/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-kube-api-access-ln2hs\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn\" (UID: \"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.195732 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn\" (UID: \"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.196123 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn\" (UID: \"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.298327 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ln2hs\" (UniqueName: \"kubernetes.io/projected/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-kube-api-access-ln2hs\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn\" (UID: \"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.298455 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn\" (UID: \"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.298574 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn\" (UID: \"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.298632 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn\" (UID: \"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.304301 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-ceph\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn\" (UID: \"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.305266 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-ssh-key\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn\" (UID: \"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.305972 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-inventory\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn\" (UID: \"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.331036 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ln2hs\" (UniqueName: \"kubernetes.io/projected/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-kube-api-access-ln2hs\") pod \"ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn\" (UID: \"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952\") " pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.412721 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn" Dec 09 16:09:41 crc kubenswrapper[4894]: I1209 16:09:41.993880 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn"] Dec 09 16:09:43 crc kubenswrapper[4894]: I1209 16:09:43.001221 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn" event={"ID":"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952","Type":"ContainerStarted","Data":"c4b18c5156d3bf1091d2d7efa9e3f417da524e29394a6aed56a556752571c560"} Dec 09 16:09:43 crc kubenswrapper[4894]: I1209 16:09:43.001722 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn" event={"ID":"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952","Type":"ContainerStarted","Data":"3da2ee8fd54996434402903b3f2ff820a0a6af2be876f90e6dda02301ec96ab1"} Dec 09 16:09:43 crc kubenswrapper[4894]: I1209 16:09:43.020884 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn" podStartSLOduration=1.511341455 podStartE2EDuration="2.020860297s" podCreationTimestamp="2025-12-09 16:09:41 +0000 UTC" firstStartedPulling="2025-12-09 16:09:42.002311536 +0000 UTC m=+2276.321522225" lastFinishedPulling="2025-12-09 16:09:42.511830398 +0000 UTC m=+2276.831041067" observedRunningTime="2025-12-09 16:09:43.018129302 +0000 UTC m=+2277.337339991" watchObservedRunningTime="2025-12-09 16:09:43.020860297 +0000 UTC m=+2277.340070986" Dec 09 16:09:47 crc kubenswrapper[4894]: I1209 16:09:47.038837 4894 generic.go:334] "Generic (PLEG): container finished" podID="8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952" containerID="c4b18c5156d3bf1091d2d7efa9e3f417da524e29394a6aed56a556752571c560" exitCode=0 Dec 09 16:09:47 crc kubenswrapper[4894]: I1209 16:09:47.038923 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn" event={"ID":"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952","Type":"ContainerDied","Data":"c4b18c5156d3bf1091d2d7efa9e3f417da524e29394a6aed56a556752571c560"} Dec 09 16:09:48 crc kubenswrapper[4894]: I1209 16:09:48.413667 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn" Dec 09 16:09:48 crc kubenswrapper[4894]: I1209 16:09:48.529579 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ln2hs\" (UniqueName: \"kubernetes.io/projected/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-kube-api-access-ln2hs\") pod \"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952\" (UID: \"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952\") " Dec 09 16:09:48 crc kubenswrapper[4894]: I1209 16:09:48.529625 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-ceph\") pod \"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952\" (UID: \"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952\") " Dec 09 16:09:48 crc kubenswrapper[4894]: I1209 16:09:48.530953 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-inventory\") pod \"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952\" (UID: \"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952\") " Dec 09 16:09:48 crc kubenswrapper[4894]: I1209 16:09:48.531011 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-ssh-key\") pod \"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952\" (UID: \"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952\") " Dec 09 16:09:48 crc kubenswrapper[4894]: I1209 16:09:48.534722 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-ceph" (OuterVolumeSpecName: "ceph") pod "8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952" (UID: "8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:09:48 crc kubenswrapper[4894]: I1209 16:09:48.536021 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-kube-api-access-ln2hs" (OuterVolumeSpecName: "kube-api-access-ln2hs") pod "8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952" (UID: "8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952"). InnerVolumeSpecName "kube-api-access-ln2hs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:09:48 crc kubenswrapper[4894]: I1209 16:09:48.552959 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-inventory" (OuterVolumeSpecName: "inventory") pod "8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952" (UID: "8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:09:48 crc kubenswrapper[4894]: I1209 16:09:48.557709 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952" (UID: "8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:09:48 crc kubenswrapper[4894]: I1209 16:09:48.633117 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 16:09:48 crc kubenswrapper[4894]: I1209 16:09:48.633391 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:09:48 crc kubenswrapper[4894]: I1209 16:09:48.633404 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ln2hs\" (UniqueName: \"kubernetes.io/projected/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-kube-api-access-ln2hs\") on node \"crc\" DevicePath \"\"" Dec 09 16:09:48 crc kubenswrapper[4894]: I1209 16:09:48.633417 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.054685 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn" event={"ID":"8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952","Type":"ContainerDied","Data":"3da2ee8fd54996434402903b3f2ff820a0a6af2be876f90e6dda02301ec96ab1"} Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.054734 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3da2ee8fd54996434402903b3f2ff820a0a6af2be876f90e6dda02301ec96ab1" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.054712 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.121746 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544"] Dec 09 16:09:49 crc kubenswrapper[4894]: E1209 16:09:49.122178 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.122200 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.122424 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952" containerName="ceph-hci-pre-edpm-deployment-openstack-edpm-ipam" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.123196 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.128939 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.129102 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.129574 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.129825 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.130016 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.133966 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544"] Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.242586 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28721fcd-587f-483d-9d67-5a5fd25a1053-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6z544\" (UID: \"28721fcd-587f-483d-9d67-5a5fd25a1053\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.242659 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28721fcd-587f-483d-9d67-5a5fd25a1053-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6z544\" (UID: \"28721fcd-587f-483d-9d67-5a5fd25a1053\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.242885 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/28721fcd-587f-483d-9d67-5a5fd25a1053-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6z544\" (UID: \"28721fcd-587f-483d-9d67-5a5fd25a1053\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.242962 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2d8cj\" (UniqueName: \"kubernetes.io/projected/28721fcd-587f-483d-9d67-5a5fd25a1053-kube-api-access-2d8cj\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6z544\" (UID: \"28721fcd-587f-483d-9d67-5a5fd25a1053\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.344400 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28721fcd-587f-483d-9d67-5a5fd25a1053-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6z544\" (UID: \"28721fcd-587f-483d-9d67-5a5fd25a1053\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.344470 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28721fcd-587f-483d-9d67-5a5fd25a1053-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6z544\" (UID: \"28721fcd-587f-483d-9d67-5a5fd25a1053\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.344589 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/28721fcd-587f-483d-9d67-5a5fd25a1053-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6z544\" (UID: \"28721fcd-587f-483d-9d67-5a5fd25a1053\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.344673 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2d8cj\" (UniqueName: \"kubernetes.io/projected/28721fcd-587f-483d-9d67-5a5fd25a1053-kube-api-access-2d8cj\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6z544\" (UID: \"28721fcd-587f-483d-9d67-5a5fd25a1053\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.349893 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28721fcd-587f-483d-9d67-5a5fd25a1053-ssh-key\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6z544\" (UID: \"28721fcd-587f-483d-9d67-5a5fd25a1053\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.350190 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/28721fcd-587f-483d-9d67-5a5fd25a1053-ceph\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6z544\" (UID: \"28721fcd-587f-483d-9d67-5a5fd25a1053\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.350222 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28721fcd-587f-483d-9d67-5a5fd25a1053-inventory\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6z544\" (UID: \"28721fcd-587f-483d-9d67-5a5fd25a1053\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.362964 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2d8cj\" (UniqueName: \"kubernetes.io/projected/28721fcd-587f-483d-9d67-5a5fd25a1053-kube-api-access-2d8cj\") pod \"configure-os-edpm-deployment-openstack-edpm-ipam-6z544\" (UID: \"28721fcd-587f-483d-9d67-5a5fd25a1053\") " pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.439794 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544" Dec 09 16:09:49 crc kubenswrapper[4894]: I1209 16:09:49.937753 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544"] Dec 09 16:09:50 crc kubenswrapper[4894]: I1209 16:09:50.063604 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544" event={"ID":"28721fcd-587f-483d-9d67-5a5fd25a1053","Type":"ContainerStarted","Data":"e9d9b6c43d2c58ac34e471134d9a13dabfedcb749665f54705fdcc7ec9ca8584"} Dec 09 16:09:51 crc kubenswrapper[4894]: I1209 16:09:51.078146 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544" event={"ID":"28721fcd-587f-483d-9d67-5a5fd25a1053","Type":"ContainerStarted","Data":"1671701f7748ed86537dfd5b01c6b22517509024fa67a89df74dffdf14d84575"} Dec 09 16:09:51 crc kubenswrapper[4894]: I1209 16:09:51.099243 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544" podStartSLOduration=1.6329616279999999 podStartE2EDuration="2.099227554s" podCreationTimestamp="2025-12-09 16:09:49 +0000 UTC" firstStartedPulling="2025-12-09 16:09:49.950254718 +0000 UTC m=+2284.269465387" lastFinishedPulling="2025-12-09 16:09:50.416520644 +0000 UTC m=+2284.735731313" observedRunningTime="2025-12-09 16:09:51.095662087 +0000 UTC m=+2285.414872776" watchObservedRunningTime="2025-12-09 16:09:51.099227554 +0000 UTC m=+2285.418438223" Dec 09 16:09:52 crc kubenswrapper[4894]: I1209 16:09:52.107508 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:09:52 crc kubenswrapper[4894]: E1209 16:09:52.108063 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:10:07 crc kubenswrapper[4894]: I1209 16:10:07.106207 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:10:07 crc kubenswrapper[4894]: E1209 16:10:07.107008 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:10:20 crc kubenswrapper[4894]: I1209 16:10:20.107117 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:10:20 crc kubenswrapper[4894]: E1209 16:10:20.107897 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:10:32 crc kubenswrapper[4894]: E1209 16:10:32.030021 4894 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28721fcd_587f_483d_9d67_5a5fd25a1053.slice/crio-conmon-1671701f7748ed86537dfd5b01c6b22517509024fa67a89df74dffdf14d84575.scope\": RecentStats: unable to find data in memory cache]" Dec 09 16:10:32 crc kubenswrapper[4894]: I1209 16:10:32.458782 4894 generic.go:334] "Generic (PLEG): container finished" podID="28721fcd-587f-483d-9d67-5a5fd25a1053" containerID="1671701f7748ed86537dfd5b01c6b22517509024fa67a89df74dffdf14d84575" exitCode=0 Dec 09 16:10:32 crc kubenswrapper[4894]: I1209 16:10:32.458836 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544" event={"ID":"28721fcd-587f-483d-9d67-5a5fd25a1053","Type":"ContainerDied","Data":"1671701f7748ed86537dfd5b01c6b22517509024fa67a89df74dffdf14d84575"} Dec 09 16:10:33 crc kubenswrapper[4894]: I1209 16:10:33.984579 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.106835 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:10:34 crc kubenswrapper[4894]: E1209 16:10:34.107169 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.133601 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28721fcd-587f-483d-9d67-5a5fd25a1053-inventory\") pod \"28721fcd-587f-483d-9d67-5a5fd25a1053\" (UID: \"28721fcd-587f-483d-9d67-5a5fd25a1053\") " Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.133752 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d8cj\" (UniqueName: \"kubernetes.io/projected/28721fcd-587f-483d-9d67-5a5fd25a1053-kube-api-access-2d8cj\") pod \"28721fcd-587f-483d-9d67-5a5fd25a1053\" (UID: \"28721fcd-587f-483d-9d67-5a5fd25a1053\") " Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.133870 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/28721fcd-587f-483d-9d67-5a5fd25a1053-ceph\") pod \"28721fcd-587f-483d-9d67-5a5fd25a1053\" (UID: \"28721fcd-587f-483d-9d67-5a5fd25a1053\") " Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.133990 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28721fcd-587f-483d-9d67-5a5fd25a1053-ssh-key\") pod \"28721fcd-587f-483d-9d67-5a5fd25a1053\" (UID: \"28721fcd-587f-483d-9d67-5a5fd25a1053\") " Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.139981 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28721fcd-587f-483d-9d67-5a5fd25a1053-ceph" (OuterVolumeSpecName: "ceph") pod "28721fcd-587f-483d-9d67-5a5fd25a1053" (UID: "28721fcd-587f-483d-9d67-5a5fd25a1053"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.140453 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28721fcd-587f-483d-9d67-5a5fd25a1053-kube-api-access-2d8cj" (OuterVolumeSpecName: "kube-api-access-2d8cj") pod "28721fcd-587f-483d-9d67-5a5fd25a1053" (UID: "28721fcd-587f-483d-9d67-5a5fd25a1053"). InnerVolumeSpecName "kube-api-access-2d8cj". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.172241 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28721fcd-587f-483d-9d67-5a5fd25a1053-inventory" (OuterVolumeSpecName: "inventory") pod "28721fcd-587f-483d-9d67-5a5fd25a1053" (UID: "28721fcd-587f-483d-9d67-5a5fd25a1053"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.178077 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28721fcd-587f-483d-9d67-5a5fd25a1053-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "28721fcd-587f-483d-9d67-5a5fd25a1053" (UID: "28721fcd-587f-483d-9d67-5a5fd25a1053"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.236214 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/28721fcd-587f-483d-9d67-5a5fd25a1053-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.236255 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d8cj\" (UniqueName: \"kubernetes.io/projected/28721fcd-587f-483d-9d67-5a5fd25a1053-kube-api-access-2d8cj\") on node \"crc\" DevicePath \"\"" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.236267 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/28721fcd-587f-483d-9d67-5a5fd25a1053-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.236277 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/28721fcd-587f-483d-9d67-5a5fd25a1053-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.478723 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544" event={"ID":"28721fcd-587f-483d-9d67-5a5fd25a1053","Type":"ContainerDied","Data":"e9d9b6c43d2c58ac34e471134d9a13dabfedcb749665f54705fdcc7ec9ca8584"} Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.479097 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e9d9b6c43d2c58ac34e471134d9a13dabfedcb749665f54705fdcc7ec9ca8584" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.478852 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/configure-os-edpm-deployment-openstack-edpm-ipam-6z544" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.599076 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-mxmp5"] Dec 09 16:10:34 crc kubenswrapper[4894]: E1209 16:10:34.599527 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28721fcd-587f-483d-9d67-5a5fd25a1053" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.599551 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="28721fcd-587f-483d-9d67-5a5fd25a1053" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.599846 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="28721fcd-587f-483d-9d67-5a5fd25a1053" containerName="configure-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.600581 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mxmp5" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.603269 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.603961 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.603990 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.604281 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.609163 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.640327 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-mxmp5"] Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.745344 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-grkjf\" (UniqueName: \"kubernetes.io/projected/797ea3d2-8204-4940-95a7-509ab68ac7a9-kube-api-access-grkjf\") pod \"ssh-known-hosts-edpm-deployment-mxmp5\" (UID: \"797ea3d2-8204-4940-95a7-509ab68ac7a9\") " pod="openstack/ssh-known-hosts-edpm-deployment-mxmp5" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.745535 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/797ea3d2-8204-4940-95a7-509ab68ac7a9-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-mxmp5\" (UID: \"797ea3d2-8204-4940-95a7-509ab68ac7a9\") " pod="openstack/ssh-known-hosts-edpm-deployment-mxmp5" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.745601 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/797ea3d2-8204-4940-95a7-509ab68ac7a9-ceph\") pod \"ssh-known-hosts-edpm-deployment-mxmp5\" (UID: \"797ea3d2-8204-4940-95a7-509ab68ac7a9\") " pod="openstack/ssh-known-hosts-edpm-deployment-mxmp5" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.745684 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/797ea3d2-8204-4940-95a7-509ab68ac7a9-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-mxmp5\" (UID: \"797ea3d2-8204-4940-95a7-509ab68ac7a9\") " pod="openstack/ssh-known-hosts-edpm-deployment-mxmp5" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.847917 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/797ea3d2-8204-4940-95a7-509ab68ac7a9-ceph\") pod \"ssh-known-hosts-edpm-deployment-mxmp5\" (UID: \"797ea3d2-8204-4940-95a7-509ab68ac7a9\") " pod="openstack/ssh-known-hosts-edpm-deployment-mxmp5" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.847995 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/797ea3d2-8204-4940-95a7-509ab68ac7a9-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-mxmp5\" (UID: \"797ea3d2-8204-4940-95a7-509ab68ac7a9\") " pod="openstack/ssh-known-hosts-edpm-deployment-mxmp5" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.848158 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-grkjf\" (UniqueName: \"kubernetes.io/projected/797ea3d2-8204-4940-95a7-509ab68ac7a9-kube-api-access-grkjf\") pod \"ssh-known-hosts-edpm-deployment-mxmp5\" (UID: \"797ea3d2-8204-4940-95a7-509ab68ac7a9\") " pod="openstack/ssh-known-hosts-edpm-deployment-mxmp5" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.848309 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/797ea3d2-8204-4940-95a7-509ab68ac7a9-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-mxmp5\" (UID: \"797ea3d2-8204-4940-95a7-509ab68ac7a9\") " pod="openstack/ssh-known-hosts-edpm-deployment-mxmp5" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.853916 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/797ea3d2-8204-4940-95a7-509ab68ac7a9-ssh-key-openstack-edpm-ipam\") pod \"ssh-known-hosts-edpm-deployment-mxmp5\" (UID: \"797ea3d2-8204-4940-95a7-509ab68ac7a9\") " pod="openstack/ssh-known-hosts-edpm-deployment-mxmp5" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.853990 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/797ea3d2-8204-4940-95a7-509ab68ac7a9-inventory-0\") pod \"ssh-known-hosts-edpm-deployment-mxmp5\" (UID: \"797ea3d2-8204-4940-95a7-509ab68ac7a9\") " pod="openstack/ssh-known-hosts-edpm-deployment-mxmp5" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.856276 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/797ea3d2-8204-4940-95a7-509ab68ac7a9-ceph\") pod \"ssh-known-hosts-edpm-deployment-mxmp5\" (UID: \"797ea3d2-8204-4940-95a7-509ab68ac7a9\") " pod="openstack/ssh-known-hosts-edpm-deployment-mxmp5" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.878669 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-grkjf\" (UniqueName: \"kubernetes.io/projected/797ea3d2-8204-4940-95a7-509ab68ac7a9-kube-api-access-grkjf\") pod \"ssh-known-hosts-edpm-deployment-mxmp5\" (UID: \"797ea3d2-8204-4940-95a7-509ab68ac7a9\") " pod="openstack/ssh-known-hosts-edpm-deployment-mxmp5" Dec 09 16:10:34 crc kubenswrapper[4894]: I1209 16:10:34.929378 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mxmp5" Dec 09 16:10:35 crc kubenswrapper[4894]: I1209 16:10:35.524529 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ssh-known-hosts-edpm-deployment-mxmp5"] Dec 09 16:10:36 crc kubenswrapper[4894]: I1209 16:10:36.495259 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mxmp5" event={"ID":"797ea3d2-8204-4940-95a7-509ab68ac7a9","Type":"ContainerStarted","Data":"6505358b32adc1b5aaf826a2ea0d250ba79798a0e29c25c26a1091ba91d4dad6"} Dec 09 16:10:36 crc kubenswrapper[4894]: I1209 16:10:36.495815 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mxmp5" event={"ID":"797ea3d2-8204-4940-95a7-509ab68ac7a9","Type":"ContainerStarted","Data":"8fdac87d6ee1494167bab27ba12a03feea4c9347911ee403616b3f571e998585"} Dec 09 16:10:36 crc kubenswrapper[4894]: I1209 16:10:36.516147 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ssh-known-hosts-edpm-deployment-mxmp5" podStartSLOduration=1.9445618919999998 podStartE2EDuration="2.51612918s" podCreationTimestamp="2025-12-09 16:10:34 +0000 UTC" firstStartedPulling="2025-12-09 16:10:35.527976677 +0000 UTC m=+2329.847187356" lastFinishedPulling="2025-12-09 16:10:36.099543965 +0000 UTC m=+2330.418754644" observedRunningTime="2025-12-09 16:10:36.510753504 +0000 UTC m=+2330.829964213" watchObservedRunningTime="2025-12-09 16:10:36.51612918 +0000 UTC m=+2330.835339859" Dec 09 16:10:45 crc kubenswrapper[4894]: I1209 16:10:45.591153 4894 generic.go:334] "Generic (PLEG): container finished" podID="797ea3d2-8204-4940-95a7-509ab68ac7a9" containerID="6505358b32adc1b5aaf826a2ea0d250ba79798a0e29c25c26a1091ba91d4dad6" exitCode=0 Dec 09 16:10:45 crc kubenswrapper[4894]: I1209 16:10:45.591281 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mxmp5" event={"ID":"797ea3d2-8204-4940-95a7-509ab68ac7a9","Type":"ContainerDied","Data":"6505358b32adc1b5aaf826a2ea0d250ba79798a0e29c25c26a1091ba91d4dad6"} Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.000664 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mxmp5" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.102002 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-grkjf\" (UniqueName: \"kubernetes.io/projected/797ea3d2-8204-4940-95a7-509ab68ac7a9-kube-api-access-grkjf\") pod \"797ea3d2-8204-4940-95a7-509ab68ac7a9\" (UID: \"797ea3d2-8204-4940-95a7-509ab68ac7a9\") " Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.102352 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/797ea3d2-8204-4940-95a7-509ab68ac7a9-ceph\") pod \"797ea3d2-8204-4940-95a7-509ab68ac7a9\" (UID: \"797ea3d2-8204-4940-95a7-509ab68ac7a9\") " Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.102471 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/797ea3d2-8204-4940-95a7-509ab68ac7a9-ssh-key-openstack-edpm-ipam\") pod \"797ea3d2-8204-4940-95a7-509ab68ac7a9\" (UID: \"797ea3d2-8204-4940-95a7-509ab68ac7a9\") " Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.102492 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/797ea3d2-8204-4940-95a7-509ab68ac7a9-inventory-0\") pod \"797ea3d2-8204-4940-95a7-509ab68ac7a9\" (UID: \"797ea3d2-8204-4940-95a7-509ab68ac7a9\") " Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.108063 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/797ea3d2-8204-4940-95a7-509ab68ac7a9-kube-api-access-grkjf" (OuterVolumeSpecName: "kube-api-access-grkjf") pod "797ea3d2-8204-4940-95a7-509ab68ac7a9" (UID: "797ea3d2-8204-4940-95a7-509ab68ac7a9"). InnerVolumeSpecName "kube-api-access-grkjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.108454 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/797ea3d2-8204-4940-95a7-509ab68ac7a9-ceph" (OuterVolumeSpecName: "ceph") pod "797ea3d2-8204-4940-95a7-509ab68ac7a9" (UID: "797ea3d2-8204-4940-95a7-509ab68ac7a9"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.136548 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/797ea3d2-8204-4940-95a7-509ab68ac7a9-inventory-0" (OuterVolumeSpecName: "inventory-0") pod "797ea3d2-8204-4940-95a7-509ab68ac7a9" (UID: "797ea3d2-8204-4940-95a7-509ab68ac7a9"). InnerVolumeSpecName "inventory-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.152223 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/797ea3d2-8204-4940-95a7-509ab68ac7a9-ssh-key-openstack-edpm-ipam" (OuterVolumeSpecName: "ssh-key-openstack-edpm-ipam") pod "797ea3d2-8204-4940-95a7-509ab68ac7a9" (UID: "797ea3d2-8204-4940-95a7-509ab68ac7a9"). InnerVolumeSpecName "ssh-key-openstack-edpm-ipam". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.205047 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key-openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/secret/797ea3d2-8204-4940-95a7-509ab68ac7a9-ssh-key-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.205093 4894 reconciler_common.go:293] "Volume detached for volume \"inventory-0\" (UniqueName: \"kubernetes.io/secret/797ea3d2-8204-4940-95a7-509ab68ac7a9-inventory-0\") on node \"crc\" DevicePath \"\"" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.205114 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-grkjf\" (UniqueName: \"kubernetes.io/projected/797ea3d2-8204-4940-95a7-509ab68ac7a9-kube-api-access-grkjf\") on node \"crc\" DevicePath \"\"" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.205131 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/797ea3d2-8204-4940-95a7-509ab68ac7a9-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.612441 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ssh-known-hosts-edpm-deployment-mxmp5" event={"ID":"797ea3d2-8204-4940-95a7-509ab68ac7a9","Type":"ContainerDied","Data":"8fdac87d6ee1494167bab27ba12a03feea4c9347911ee403616b3f571e998585"} Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.612735 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8fdac87d6ee1494167bab27ba12a03feea4c9347911ee403616b3f571e998585" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.612518 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ssh-known-hosts-edpm-deployment-mxmp5" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.689779 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf"] Dec 09 16:10:47 crc kubenswrapper[4894]: E1209 16:10:47.690139 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="797ea3d2-8204-4940-95a7-509ab68ac7a9" containerName="ssh-known-hosts-edpm-deployment" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.690153 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="797ea3d2-8204-4940-95a7-509ab68ac7a9" containerName="ssh-known-hosts-edpm-deployment" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.690361 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="797ea3d2-8204-4940-95a7-509ab68ac7a9" containerName="ssh-known-hosts-edpm-deployment" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.691436 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.693483 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.693941 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.694053 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.694301 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.695593 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.706976 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf"] Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.815345 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c58409cd-be24-47f2-869f-6bfd530ab945-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmhtf\" (UID: \"c58409cd-be24-47f2-869f-6bfd530ab945\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.815459 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c58409cd-be24-47f2-869f-6bfd530ab945-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmhtf\" (UID: \"c58409cd-be24-47f2-869f-6bfd530ab945\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.815532 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pz54\" (UniqueName: \"kubernetes.io/projected/c58409cd-be24-47f2-869f-6bfd530ab945-kube-api-access-5pz54\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmhtf\" (UID: \"c58409cd-be24-47f2-869f-6bfd530ab945\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.815618 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c58409cd-be24-47f2-869f-6bfd530ab945-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmhtf\" (UID: \"c58409cd-be24-47f2-869f-6bfd530ab945\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.918259 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c58409cd-be24-47f2-869f-6bfd530ab945-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmhtf\" (UID: \"c58409cd-be24-47f2-869f-6bfd530ab945\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.918396 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c58409cd-be24-47f2-869f-6bfd530ab945-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmhtf\" (UID: \"c58409cd-be24-47f2-869f-6bfd530ab945\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.918463 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c58409cd-be24-47f2-869f-6bfd530ab945-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmhtf\" (UID: \"c58409cd-be24-47f2-869f-6bfd530ab945\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.918527 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pz54\" (UniqueName: \"kubernetes.io/projected/c58409cd-be24-47f2-869f-6bfd530ab945-kube-api-access-5pz54\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmhtf\" (UID: \"c58409cd-be24-47f2-869f-6bfd530ab945\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.923818 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c58409cd-be24-47f2-869f-6bfd530ab945-ceph\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmhtf\" (UID: \"c58409cd-be24-47f2-869f-6bfd530ab945\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.928212 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c58409cd-be24-47f2-869f-6bfd530ab945-ssh-key\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmhtf\" (UID: \"c58409cd-be24-47f2-869f-6bfd530ab945\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.928246 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c58409cd-be24-47f2-869f-6bfd530ab945-inventory\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmhtf\" (UID: \"c58409cd-be24-47f2-869f-6bfd530ab945\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf" Dec 09 16:10:47 crc kubenswrapper[4894]: I1209 16:10:47.944360 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pz54\" (UniqueName: \"kubernetes.io/projected/c58409cd-be24-47f2-869f-6bfd530ab945-kube-api-access-5pz54\") pod \"run-os-edpm-deployment-openstack-edpm-ipam-jmhtf\" (UID: \"c58409cd-be24-47f2-869f-6bfd530ab945\") " pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf" Dec 09 16:10:48 crc kubenswrapper[4894]: I1209 16:10:48.013533 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf" Dec 09 16:10:48 crc kubenswrapper[4894]: I1209 16:10:48.107558 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:10:48 crc kubenswrapper[4894]: E1209 16:10:48.108160 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:10:48 crc kubenswrapper[4894]: I1209 16:10:48.573726 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf"] Dec 09 16:10:48 crc kubenswrapper[4894]: I1209 16:10:48.620774 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf" event={"ID":"c58409cd-be24-47f2-869f-6bfd530ab945","Type":"ContainerStarted","Data":"7752c0dc3859ba8ba36a9c390d9bf4737fdc5965005b9098b5e9e3f18e318624"} Dec 09 16:10:49 crc kubenswrapper[4894]: I1209 16:10:49.634759 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf" event={"ID":"c58409cd-be24-47f2-869f-6bfd530ab945","Type":"ContainerStarted","Data":"4aabf78175e9303a118fd4b76ca17ec251b0a693a49de462a3e996aa684b31f0"} Dec 09 16:10:58 crc kubenswrapper[4894]: I1209 16:10:58.748726 4894 generic.go:334] "Generic (PLEG): container finished" podID="c58409cd-be24-47f2-869f-6bfd530ab945" containerID="4aabf78175e9303a118fd4b76ca17ec251b0a693a49de462a3e996aa684b31f0" exitCode=0 Dec 09 16:10:58 crc kubenswrapper[4894]: I1209 16:10:58.749555 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf" event={"ID":"c58409cd-be24-47f2-869f-6bfd530ab945","Type":"ContainerDied","Data":"4aabf78175e9303a118fd4b76ca17ec251b0a693a49de462a3e996aa684b31f0"} Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.197580 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf" Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.287604 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c58409cd-be24-47f2-869f-6bfd530ab945-ssh-key\") pod \"c58409cd-be24-47f2-869f-6bfd530ab945\" (UID: \"c58409cd-be24-47f2-869f-6bfd530ab945\") " Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.287707 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c58409cd-be24-47f2-869f-6bfd530ab945-inventory\") pod \"c58409cd-be24-47f2-869f-6bfd530ab945\" (UID: \"c58409cd-be24-47f2-869f-6bfd530ab945\") " Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.287744 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5pz54\" (UniqueName: \"kubernetes.io/projected/c58409cd-be24-47f2-869f-6bfd530ab945-kube-api-access-5pz54\") pod \"c58409cd-be24-47f2-869f-6bfd530ab945\" (UID: \"c58409cd-be24-47f2-869f-6bfd530ab945\") " Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.287803 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c58409cd-be24-47f2-869f-6bfd530ab945-ceph\") pod \"c58409cd-be24-47f2-869f-6bfd530ab945\" (UID: \"c58409cd-be24-47f2-869f-6bfd530ab945\") " Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.293943 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c58409cd-be24-47f2-869f-6bfd530ab945-kube-api-access-5pz54" (OuterVolumeSpecName: "kube-api-access-5pz54") pod "c58409cd-be24-47f2-869f-6bfd530ab945" (UID: "c58409cd-be24-47f2-869f-6bfd530ab945"). InnerVolumeSpecName "kube-api-access-5pz54". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.294961 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c58409cd-be24-47f2-869f-6bfd530ab945-ceph" (OuterVolumeSpecName: "ceph") pod "c58409cd-be24-47f2-869f-6bfd530ab945" (UID: "c58409cd-be24-47f2-869f-6bfd530ab945"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.317904 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c58409cd-be24-47f2-869f-6bfd530ab945-inventory" (OuterVolumeSpecName: "inventory") pod "c58409cd-be24-47f2-869f-6bfd530ab945" (UID: "c58409cd-be24-47f2-869f-6bfd530ab945"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.340983 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c58409cd-be24-47f2-869f-6bfd530ab945-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "c58409cd-be24-47f2-869f-6bfd530ab945" (UID: "c58409cd-be24-47f2-869f-6bfd530ab945"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.390588 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5pz54\" (UniqueName: \"kubernetes.io/projected/c58409cd-be24-47f2-869f-6bfd530ab945-kube-api-access-5pz54\") on node \"crc\" DevicePath \"\"" Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.390792 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/c58409cd-be24-47f2-869f-6bfd530ab945-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.390865 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/c58409cd-be24-47f2-869f-6bfd530ab945-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.390932 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/c58409cd-be24-47f2-869f-6bfd530ab945-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.774575 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf" event={"ID":"c58409cd-be24-47f2-869f-6bfd530ab945","Type":"ContainerDied","Data":"7752c0dc3859ba8ba36a9c390d9bf4737fdc5965005b9098b5e9e3f18e318624"} Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.774652 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7752c0dc3859ba8ba36a9c390d9bf4737fdc5965005b9098b5e9e3f18e318624" Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.774639 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/run-os-edpm-deployment-openstack-edpm-ipam-jmhtf" Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.884682 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t"] Dec 09 16:11:00 crc kubenswrapper[4894]: E1209 16:11:00.885241 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c58409cd-be24-47f2-869f-6bfd530ab945" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.885279 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c58409cd-be24-47f2-869f-6bfd530ab945" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.885551 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="c58409cd-be24-47f2-869f-6bfd530ab945" containerName="run-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.886284 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t" Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.888163 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.888879 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.889015 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.889041 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.890102 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 16:11:00 crc kubenswrapper[4894]: I1209 16:11:00.920881 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t"] Dec 09 16:11:01 crc kubenswrapper[4894]: I1209 16:11:01.009193 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b6715a4-d866-40d0-8e01-40d603a65b79-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t\" (UID: \"1b6715a4-d866-40d0-8e01-40d603a65b79\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t" Dec 09 16:11:01 crc kubenswrapper[4894]: I1209 16:11:01.009283 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1b6715a4-d866-40d0-8e01-40d603a65b79-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t\" (UID: \"1b6715a4-d866-40d0-8e01-40d603a65b79\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t" Dec 09 16:11:01 crc kubenswrapper[4894]: I1209 16:11:01.009352 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lz84k\" (UniqueName: \"kubernetes.io/projected/1b6715a4-d866-40d0-8e01-40d603a65b79-kube-api-access-lz84k\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t\" (UID: \"1b6715a4-d866-40d0-8e01-40d603a65b79\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t" Dec 09 16:11:01 crc kubenswrapper[4894]: I1209 16:11:01.009520 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b6715a4-d866-40d0-8e01-40d603a65b79-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t\" (UID: \"1b6715a4-d866-40d0-8e01-40d603a65b79\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t" Dec 09 16:11:01 crc kubenswrapper[4894]: I1209 16:11:01.111115 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b6715a4-d866-40d0-8e01-40d603a65b79-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t\" (UID: \"1b6715a4-d866-40d0-8e01-40d603a65b79\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t" Dec 09 16:11:01 crc kubenswrapper[4894]: I1209 16:11:01.111474 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1b6715a4-d866-40d0-8e01-40d603a65b79-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t\" (UID: \"1b6715a4-d866-40d0-8e01-40d603a65b79\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t" Dec 09 16:11:01 crc kubenswrapper[4894]: I1209 16:11:01.111590 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lz84k\" (UniqueName: \"kubernetes.io/projected/1b6715a4-d866-40d0-8e01-40d603a65b79-kube-api-access-lz84k\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t\" (UID: \"1b6715a4-d866-40d0-8e01-40d603a65b79\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t" Dec 09 16:11:01 crc kubenswrapper[4894]: I1209 16:11:01.111802 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b6715a4-d866-40d0-8e01-40d603a65b79-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t\" (UID: \"1b6715a4-d866-40d0-8e01-40d603a65b79\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t" Dec 09 16:11:01 crc kubenswrapper[4894]: I1209 16:11:01.115113 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b6715a4-d866-40d0-8e01-40d603a65b79-ssh-key\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t\" (UID: \"1b6715a4-d866-40d0-8e01-40d603a65b79\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t" Dec 09 16:11:01 crc kubenswrapper[4894]: I1209 16:11:01.119161 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1b6715a4-d866-40d0-8e01-40d603a65b79-ceph\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t\" (UID: \"1b6715a4-d866-40d0-8e01-40d603a65b79\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t" Dec 09 16:11:01 crc kubenswrapper[4894]: I1209 16:11:01.120820 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b6715a4-d866-40d0-8e01-40d603a65b79-inventory\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t\" (UID: \"1b6715a4-d866-40d0-8e01-40d603a65b79\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t" Dec 09 16:11:01 crc kubenswrapper[4894]: I1209 16:11:01.130200 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lz84k\" (UniqueName: \"kubernetes.io/projected/1b6715a4-d866-40d0-8e01-40d603a65b79-kube-api-access-lz84k\") pod \"reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t\" (UID: \"1b6715a4-d866-40d0-8e01-40d603a65b79\") " pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t" Dec 09 16:11:01 crc kubenswrapper[4894]: I1209 16:11:01.225665 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t" Dec 09 16:11:01 crc kubenswrapper[4894]: I1209 16:11:01.750274 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t"] Dec 09 16:11:01 crc kubenswrapper[4894]: I1209 16:11:01.783500 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t" event={"ID":"1b6715a4-d866-40d0-8e01-40d603a65b79","Type":"ContainerStarted","Data":"c467e372f22555b505bd56fc2b0df1c4c57b22938a20a81ae834bfaee50c723c"} Dec 09 16:11:02 crc kubenswrapper[4894]: I1209 16:11:02.107010 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:11:02 crc kubenswrapper[4894]: E1209 16:11:02.107469 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:11:03 crc kubenswrapper[4894]: I1209 16:11:03.807224 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t" event={"ID":"1b6715a4-d866-40d0-8e01-40d603a65b79","Type":"ContainerStarted","Data":"a94fa8507135cdf1469ecf2408f8894213ac2764863cc991564d097ac7fc4a31"} Dec 09 16:11:03 crc kubenswrapper[4894]: I1209 16:11:03.836559 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t" podStartSLOduration=2.785049591 podStartE2EDuration="3.836533637s" podCreationTimestamp="2025-12-09 16:11:00 +0000 UTC" firstStartedPulling="2025-12-09 16:11:01.759307816 +0000 UTC m=+2356.078518495" lastFinishedPulling="2025-12-09 16:11:02.810791862 +0000 UTC m=+2357.130002541" observedRunningTime="2025-12-09 16:11:03.830882543 +0000 UTC m=+2358.150093282" watchObservedRunningTime="2025-12-09 16:11:03.836533637 +0000 UTC m=+2358.155744336" Dec 09 16:11:12 crc kubenswrapper[4894]: I1209 16:11:12.895221 4894 generic.go:334] "Generic (PLEG): container finished" podID="1b6715a4-d866-40d0-8e01-40d603a65b79" containerID="a94fa8507135cdf1469ecf2408f8894213ac2764863cc991564d097ac7fc4a31" exitCode=0 Dec 09 16:11:12 crc kubenswrapper[4894]: I1209 16:11:12.895287 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t" event={"ID":"1b6715a4-d866-40d0-8e01-40d603a65b79","Type":"ContainerDied","Data":"a94fa8507135cdf1469ecf2408f8894213ac2764863cc991564d097ac7fc4a31"} Dec 09 16:11:14 crc kubenswrapper[4894]: I1209 16:11:14.317573 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t" Dec 09 16:11:14 crc kubenswrapper[4894]: I1209 16:11:14.510853 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz84k\" (UniqueName: \"kubernetes.io/projected/1b6715a4-d866-40d0-8e01-40d603a65b79-kube-api-access-lz84k\") pod \"1b6715a4-d866-40d0-8e01-40d603a65b79\" (UID: \"1b6715a4-d866-40d0-8e01-40d603a65b79\") " Dec 09 16:11:14 crc kubenswrapper[4894]: I1209 16:11:14.510999 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b6715a4-d866-40d0-8e01-40d603a65b79-ssh-key\") pod \"1b6715a4-d866-40d0-8e01-40d603a65b79\" (UID: \"1b6715a4-d866-40d0-8e01-40d603a65b79\") " Dec 09 16:11:14 crc kubenswrapper[4894]: I1209 16:11:14.511055 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1b6715a4-d866-40d0-8e01-40d603a65b79-ceph\") pod \"1b6715a4-d866-40d0-8e01-40d603a65b79\" (UID: \"1b6715a4-d866-40d0-8e01-40d603a65b79\") " Dec 09 16:11:14 crc kubenswrapper[4894]: I1209 16:11:14.511151 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b6715a4-d866-40d0-8e01-40d603a65b79-inventory\") pod \"1b6715a4-d866-40d0-8e01-40d603a65b79\" (UID: \"1b6715a4-d866-40d0-8e01-40d603a65b79\") " Dec 09 16:11:14 crc kubenswrapper[4894]: I1209 16:11:14.517801 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b6715a4-d866-40d0-8e01-40d603a65b79-ceph" (OuterVolumeSpecName: "ceph") pod "1b6715a4-d866-40d0-8e01-40d603a65b79" (UID: "1b6715a4-d866-40d0-8e01-40d603a65b79"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:11:14 crc kubenswrapper[4894]: I1209 16:11:14.517944 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1b6715a4-d866-40d0-8e01-40d603a65b79-kube-api-access-lz84k" (OuterVolumeSpecName: "kube-api-access-lz84k") pod "1b6715a4-d866-40d0-8e01-40d603a65b79" (UID: "1b6715a4-d866-40d0-8e01-40d603a65b79"). InnerVolumeSpecName "kube-api-access-lz84k". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:11:14 crc kubenswrapper[4894]: I1209 16:11:14.545491 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b6715a4-d866-40d0-8e01-40d603a65b79-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "1b6715a4-d866-40d0-8e01-40d603a65b79" (UID: "1b6715a4-d866-40d0-8e01-40d603a65b79"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:11:14 crc kubenswrapper[4894]: I1209 16:11:14.562549 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1b6715a4-d866-40d0-8e01-40d603a65b79-inventory" (OuterVolumeSpecName: "inventory") pod "1b6715a4-d866-40d0-8e01-40d603a65b79" (UID: "1b6715a4-d866-40d0-8e01-40d603a65b79"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:11:14 crc kubenswrapper[4894]: I1209 16:11:14.616383 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz84k\" (UniqueName: \"kubernetes.io/projected/1b6715a4-d866-40d0-8e01-40d603a65b79-kube-api-access-lz84k\") on node \"crc\" DevicePath \"\"" Dec 09 16:11:14 crc kubenswrapper[4894]: I1209 16:11:14.616429 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/1b6715a4-d866-40d0-8e01-40d603a65b79-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:11:14 crc kubenswrapper[4894]: I1209 16:11:14.616442 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/1b6715a4-d866-40d0-8e01-40d603a65b79-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:11:14 crc kubenswrapper[4894]: I1209 16:11:14.616455 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/1b6715a4-d866-40d0-8e01-40d603a65b79-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 16:11:14 crc kubenswrapper[4894]: I1209 16:11:14.932954 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t" Dec 09 16:11:14 crc kubenswrapper[4894]: I1209 16:11:14.935761 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t" event={"ID":"1b6715a4-d866-40d0-8e01-40d603a65b79","Type":"ContainerDied","Data":"c467e372f22555b505bd56fc2b0df1c4c57b22938a20a81ae834bfaee50c723c"} Dec 09 16:11:14 crc kubenswrapper[4894]: I1209 16:11:14.935865 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c467e372f22555b505bd56fc2b0df1c4c57b22938a20a81ae834bfaee50c723c" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.022715 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv"] Dec 09 16:11:15 crc kubenswrapper[4894]: E1209 16:11:15.023397 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1b6715a4-d866-40d0-8e01-40d603a65b79" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.023428 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="1b6715a4-d866-40d0-8e01-40d603a65b79" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.023747 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="1b6715a4-d866-40d0-8e01-40d603a65b79" containerName="reboot-os-edpm-deployment-openstack-edpm-ipam" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.024621 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.027832 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-libvirt-default-certs-0" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.028146 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-ovn-default-certs-0" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.028345 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-neutron-metadata-default-certs-0" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.028581 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.028806 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.029298 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.029485 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.035801 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv"] Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.039749 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.228920 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.228972 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.228994 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.229013 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.229141 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p48dr\" (UniqueName: \"kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-kube-api-access-p48dr\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.229806 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.230051 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.230276 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.230382 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.230422 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.230450 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.230566 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.230731 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.332343 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.332434 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.332502 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.332552 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.332595 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.332663 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.332714 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.332789 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.332892 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.332940 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.332967 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.333000 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.333037 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p48dr\" (UniqueName: \"kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-kube-api-access-p48dr\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.338090 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-ssh-key\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.338386 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-ovn-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.338620 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.338976 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-ceph\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.339016 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-openstack-edpm-ipam-ovn-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.339773 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-repo-setup-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.340043 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-libvirt-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.340363 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-bootstrap-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.340600 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-neutron-metadata-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.341603 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.342191 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-inventory\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.342386 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-nova-combined-ca-bundle\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.358872 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p48dr\" (UniqueName: \"kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-kube-api-access-p48dr\") pod \"install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:15 crc kubenswrapper[4894]: I1209 16:11:15.657619 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:16 crc kubenswrapper[4894]: W1209 16:11:16.169076 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3c2da51_8857_42f2_b4a4_731927ed1599.slice/crio-3dbf7a1cac127644c2b8c1919f5c4ebf6628ca81ba94788c8655361c38c6c834 WatchSource:0}: Error finding container 3dbf7a1cac127644c2b8c1919f5c4ebf6628ca81ba94788c8655361c38c6c834: Status 404 returned error can't find the container with id 3dbf7a1cac127644c2b8c1919f5c4ebf6628ca81ba94788c8655361c38c6c834 Dec 09 16:11:16 crc kubenswrapper[4894]: I1209 16:11:16.170882 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv"] Dec 09 16:11:16 crc kubenswrapper[4894]: I1209 16:11:16.948306 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" event={"ID":"f3c2da51-8857-42f2-b4a4-731927ed1599","Type":"ContainerStarted","Data":"3dbf7a1cac127644c2b8c1919f5c4ebf6628ca81ba94788c8655361c38c6c834"} Dec 09 16:11:17 crc kubenswrapper[4894]: I1209 16:11:17.107473 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:11:17 crc kubenswrapper[4894]: E1209 16:11:17.108384 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:11:17 crc kubenswrapper[4894]: I1209 16:11:17.963225 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" event={"ID":"f3c2da51-8857-42f2-b4a4-731927ed1599","Type":"ContainerStarted","Data":"9e4933a095f84e9be907942d70e5e57abb15ea914d21e8b6f3e72ef06f1561a2"} Dec 09 16:11:18 crc kubenswrapper[4894]: I1209 16:11:18.002488 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" podStartSLOduration=3.341911042 podStartE2EDuration="4.00247214s" podCreationTimestamp="2025-12-09 16:11:14 +0000 UTC" firstStartedPulling="2025-12-09 16:11:16.171611386 +0000 UTC m=+2370.490822055" lastFinishedPulling="2025-12-09 16:11:16.832172484 +0000 UTC m=+2371.151383153" observedRunningTime="2025-12-09 16:11:17.997478394 +0000 UTC m=+2372.316689063" watchObservedRunningTime="2025-12-09 16:11:18.00247214 +0000 UTC m=+2372.321682809" Dec 09 16:11:31 crc kubenswrapper[4894]: I1209 16:11:31.107052 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:11:31 crc kubenswrapper[4894]: E1209 16:11:31.107837 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:11:45 crc kubenswrapper[4894]: I1209 16:11:45.106743 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:11:45 crc kubenswrapper[4894]: E1209 16:11:45.107476 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:11:50 crc kubenswrapper[4894]: I1209 16:11:50.269954 4894 generic.go:334] "Generic (PLEG): container finished" podID="f3c2da51-8857-42f2-b4a4-731927ed1599" containerID="9e4933a095f84e9be907942d70e5e57abb15ea914d21e8b6f3e72ef06f1561a2" exitCode=0 Dec 09 16:11:50 crc kubenswrapper[4894]: I1209 16:11:50.270017 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" event={"ID":"f3c2da51-8857-42f2-b4a4-731927ed1599","Type":"ContainerDied","Data":"9e4933a095f84e9be907942d70e5e57abb15ea914d21e8b6f3e72ef06f1561a2"} Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.706045 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.841536 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-libvirt-combined-ca-bundle\") pod \"f3c2da51-8857-42f2-b4a4-731927ed1599\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.841618 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-repo-setup-combined-ca-bundle\") pod \"f3c2da51-8857-42f2-b4a4-731927ed1599\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.841673 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p48dr\" (UniqueName: \"kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-kube-api-access-p48dr\") pod \"f3c2da51-8857-42f2-b4a4-731927ed1599\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.841719 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-inventory\") pod \"f3c2da51-8857-42f2-b4a4-731927ed1599\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.841766 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-neutron-metadata-combined-ca-bundle\") pod \"f3c2da51-8857-42f2-b4a4-731927ed1599\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.841826 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-openstack-edpm-ipam-ovn-default-certs-0\") pod \"f3c2da51-8857-42f2-b4a4-731927ed1599\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.841940 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-openstack-edpm-ipam-libvirt-default-certs-0\") pod \"f3c2da51-8857-42f2-b4a4-731927ed1599\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.842005 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-ssh-key\") pod \"f3c2da51-8857-42f2-b4a4-731927ed1599\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.842042 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-ovn-combined-ca-bundle\") pod \"f3c2da51-8857-42f2-b4a4-731927ed1599\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.842067 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-ceph\") pod \"f3c2da51-8857-42f2-b4a4-731927ed1599\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.842113 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-bootstrap-combined-ca-bundle\") pod \"f3c2da51-8857-42f2-b4a4-731927ed1599\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.842149 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-openstack-edpm-ipam-neutron-metadata-default-certs-0\") pod \"f3c2da51-8857-42f2-b4a4-731927ed1599\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.842190 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-nova-combined-ca-bundle\") pod \"f3c2da51-8857-42f2-b4a4-731927ed1599\" (UID: \"f3c2da51-8857-42f2-b4a4-731927ed1599\") " Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.848182 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-ceph" (OuterVolumeSpecName: "ceph") pod "f3c2da51-8857-42f2-b4a4-731927ed1599" (UID: "f3c2da51-8857-42f2-b4a4-731927ed1599"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.848219 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "f3c2da51-8857-42f2-b4a4-731927ed1599" (UID: "f3c2da51-8857-42f2-b4a4-731927ed1599"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.848624 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-nova-combined-ca-bundle" (OuterVolumeSpecName: "nova-combined-ca-bundle") pod "f3c2da51-8857-42f2-b4a4-731927ed1599" (UID: "f3c2da51-8857-42f2-b4a4-731927ed1599"). InnerVolumeSpecName "nova-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.848739 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "f3c2da51-8857-42f2-b4a4-731927ed1599" (UID: "f3c2da51-8857-42f2-b4a4-731927ed1599"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.848729 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-bootstrap-combined-ca-bundle" (OuterVolumeSpecName: "bootstrap-combined-ca-bundle") pod "f3c2da51-8857-42f2-b4a4-731927ed1599" (UID: "f3c2da51-8857-42f2-b4a4-731927ed1599"). InnerVolumeSpecName "bootstrap-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.849054 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "f3c2da51-8857-42f2-b4a4-731927ed1599" (UID: "f3c2da51-8857-42f2-b4a4-731927ed1599"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.849474 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-repo-setup-combined-ca-bundle" (OuterVolumeSpecName: "repo-setup-combined-ca-bundle") pod "f3c2da51-8857-42f2-b4a4-731927ed1599" (UID: "f3c2da51-8857-42f2-b4a4-731927ed1599"). InnerVolumeSpecName "repo-setup-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.850768 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-openstack-edpm-ipam-ovn-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-ovn-default-certs-0") pod "f3c2da51-8857-42f2-b4a4-731927ed1599" (UID: "f3c2da51-8857-42f2-b4a4-731927ed1599"). InnerVolumeSpecName "openstack-edpm-ipam-ovn-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.850803 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-openstack-edpm-ipam-neutron-metadata-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-neutron-metadata-default-certs-0") pod "f3c2da51-8857-42f2-b4a4-731927ed1599" (UID: "f3c2da51-8857-42f2-b4a4-731927ed1599"). InnerVolumeSpecName "openstack-edpm-ipam-neutron-metadata-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.851531 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-kube-api-access-p48dr" (OuterVolumeSpecName: "kube-api-access-p48dr") pod "f3c2da51-8857-42f2-b4a4-731927ed1599" (UID: "f3c2da51-8857-42f2-b4a4-731927ed1599"). InnerVolumeSpecName "kube-api-access-p48dr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.851881 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-openstack-edpm-ipam-libvirt-default-certs-0" (OuterVolumeSpecName: "openstack-edpm-ipam-libvirt-default-certs-0") pod "f3c2da51-8857-42f2-b4a4-731927ed1599" (UID: "f3c2da51-8857-42f2-b4a4-731927ed1599"). InnerVolumeSpecName "openstack-edpm-ipam-libvirt-default-certs-0". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.871293 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f3c2da51-8857-42f2-b4a4-731927ed1599" (UID: "f3c2da51-8857-42f2-b4a4-731927ed1599"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.871977 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-inventory" (OuterVolumeSpecName: "inventory") pod "f3c2da51-8857-42f2-b4a4-731927ed1599" (UID: "f3c2da51-8857-42f2-b4a4-731927ed1599"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.944445 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.944477 4894 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.944491 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.944500 4894 reconciler_common.go:293] "Volume detached for volume \"bootstrap-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-bootstrap-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.944510 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-neutron-metadata-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-openstack-edpm-ipam-neutron-metadata-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.944519 4894 reconciler_common.go:293] "Volume detached for volume \"nova-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-nova-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.944527 4894 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.944536 4894 reconciler_common.go:293] "Volume detached for volume \"repo-setup-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-repo-setup-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.944545 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p48dr\" (UniqueName: \"kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-kube-api-access-p48dr\") on node \"crc\" DevicePath \"\"" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.944553 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.944561 4894 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3c2da51-8857-42f2-b4a4-731927ed1599-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.944572 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-ovn-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-openstack-edpm-ipam-ovn-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 09 16:11:51 crc kubenswrapper[4894]: I1209 16:11:51.944583 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam-libvirt-default-certs-0\" (UniqueName: \"kubernetes.io/projected/f3c2da51-8857-42f2-b4a4-731927ed1599-openstack-edpm-ipam-libvirt-default-certs-0\") on node \"crc\" DevicePath \"\"" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.291418 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" event={"ID":"f3c2da51-8857-42f2-b4a4-731927ed1599","Type":"ContainerDied","Data":"3dbf7a1cac127644c2b8c1919f5c4ebf6628ca81ba94788c8655361c38c6c834"} Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.291462 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3dbf7a1cac127644c2b8c1919f5c4ebf6628ca81ba94788c8655361c38c6c834" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.291548 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.473566 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g"] Dec 09 16:11:52 crc kubenswrapper[4894]: E1209 16:11:52.474434 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3c2da51-8857-42f2-b4a4-731927ed1599" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.474477 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3c2da51-8857-42f2-b4a4-731927ed1599" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.474986 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3c2da51-8857-42f2-b4a4-731927ed1599" containerName="install-certs-edpm-deployment-openstack-edpm-ipam" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.476330 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.488923 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g"] Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.514208 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.514321 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.515936 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.516022 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.516299 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.555547 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzhxx\" (UniqueName: \"kubernetes.io/projected/260e7406-3b55-4acf-a34a-3588a8bd1176-kube-api-access-zzhxx\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g\" (UID: \"260e7406-3b55-4acf-a34a-3588a8bd1176\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.555784 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/260e7406-3b55-4acf-a34a-3588a8bd1176-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g\" (UID: \"260e7406-3b55-4acf-a34a-3588a8bd1176\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.555977 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/260e7406-3b55-4acf-a34a-3588a8bd1176-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g\" (UID: \"260e7406-3b55-4acf-a34a-3588a8bd1176\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.556096 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/260e7406-3b55-4acf-a34a-3588a8bd1176-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g\" (UID: \"260e7406-3b55-4acf-a34a-3588a8bd1176\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.657385 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzhxx\" (UniqueName: \"kubernetes.io/projected/260e7406-3b55-4acf-a34a-3588a8bd1176-kube-api-access-zzhxx\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g\" (UID: \"260e7406-3b55-4acf-a34a-3588a8bd1176\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.657458 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/260e7406-3b55-4acf-a34a-3588a8bd1176-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g\" (UID: \"260e7406-3b55-4acf-a34a-3588a8bd1176\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.657522 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/260e7406-3b55-4acf-a34a-3588a8bd1176-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g\" (UID: \"260e7406-3b55-4acf-a34a-3588a8bd1176\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.657553 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/260e7406-3b55-4acf-a34a-3588a8bd1176-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g\" (UID: \"260e7406-3b55-4acf-a34a-3588a8bd1176\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.661009 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/260e7406-3b55-4acf-a34a-3588a8bd1176-ceph\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g\" (UID: \"260e7406-3b55-4acf-a34a-3588a8bd1176\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.661086 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/260e7406-3b55-4acf-a34a-3588a8bd1176-inventory\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g\" (UID: \"260e7406-3b55-4acf-a34a-3588a8bd1176\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.664787 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/260e7406-3b55-4acf-a34a-3588a8bd1176-ssh-key\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g\" (UID: \"260e7406-3b55-4acf-a34a-3588a8bd1176\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.687327 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzhxx\" (UniqueName: \"kubernetes.io/projected/260e7406-3b55-4acf-a34a-3588a8bd1176-kube-api-access-zzhxx\") pod \"ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g\" (UID: \"260e7406-3b55-4acf-a34a-3588a8bd1176\") " pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g" Dec 09 16:11:52 crc kubenswrapper[4894]: I1209 16:11:52.848734 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g" Dec 09 16:11:53 crc kubenswrapper[4894]: I1209 16:11:53.396951 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g"] Dec 09 16:11:54 crc kubenswrapper[4894]: I1209 16:11:54.317005 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g" event={"ID":"260e7406-3b55-4acf-a34a-3588a8bd1176","Type":"ContainerStarted","Data":"e69786697e856660a53b2235bba9e97a01f3a961a8a84f930b765ed73053afb1"} Dec 09 16:11:54 crc kubenswrapper[4894]: I1209 16:11:54.318207 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g" event={"ID":"260e7406-3b55-4acf-a34a-3588a8bd1176","Type":"ContainerStarted","Data":"1723afa30179ead6279b982636b57204e333b78ee8da000e7544800e62c212d8"} Dec 09 16:11:54 crc kubenswrapper[4894]: I1209 16:11:54.343688 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g" podStartSLOduration=1.8747448549999999 podStartE2EDuration="2.343664527s" podCreationTimestamp="2025-12-09 16:11:52 +0000 UTC" firstStartedPulling="2025-12-09 16:11:53.405688919 +0000 UTC m=+2407.724899588" lastFinishedPulling="2025-12-09 16:11:53.874608591 +0000 UTC m=+2408.193819260" observedRunningTime="2025-12-09 16:11:54.331883378 +0000 UTC m=+2408.651094047" watchObservedRunningTime="2025-12-09 16:11:54.343664527 +0000 UTC m=+2408.662875196" Dec 09 16:11:57 crc kubenswrapper[4894]: I1209 16:11:57.106339 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:11:57 crc kubenswrapper[4894]: E1209 16:11:57.108303 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:11:59 crc kubenswrapper[4894]: I1209 16:11:59.361714 4894 generic.go:334] "Generic (PLEG): container finished" podID="260e7406-3b55-4acf-a34a-3588a8bd1176" containerID="e69786697e856660a53b2235bba9e97a01f3a961a8a84f930b765ed73053afb1" exitCode=0 Dec 09 16:11:59 crc kubenswrapper[4894]: I1209 16:11:59.361803 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g" event={"ID":"260e7406-3b55-4acf-a34a-3588a8bd1176","Type":"ContainerDied","Data":"e69786697e856660a53b2235bba9e97a01f3a961a8a84f930b765ed73053afb1"} Dec 09 16:12:00 crc kubenswrapper[4894]: I1209 16:12:00.801263 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g" Dec 09 16:12:00 crc kubenswrapper[4894]: I1209 16:12:00.914144 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/260e7406-3b55-4acf-a34a-3588a8bd1176-inventory\") pod \"260e7406-3b55-4acf-a34a-3588a8bd1176\" (UID: \"260e7406-3b55-4acf-a34a-3588a8bd1176\") " Dec 09 16:12:00 crc kubenswrapper[4894]: I1209 16:12:00.914386 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzhxx\" (UniqueName: \"kubernetes.io/projected/260e7406-3b55-4acf-a34a-3588a8bd1176-kube-api-access-zzhxx\") pod \"260e7406-3b55-4acf-a34a-3588a8bd1176\" (UID: \"260e7406-3b55-4acf-a34a-3588a8bd1176\") " Dec 09 16:12:00 crc kubenswrapper[4894]: I1209 16:12:00.914475 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/260e7406-3b55-4acf-a34a-3588a8bd1176-ssh-key\") pod \"260e7406-3b55-4acf-a34a-3588a8bd1176\" (UID: \"260e7406-3b55-4acf-a34a-3588a8bd1176\") " Dec 09 16:12:00 crc kubenswrapper[4894]: I1209 16:12:00.914554 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/260e7406-3b55-4acf-a34a-3588a8bd1176-ceph\") pod \"260e7406-3b55-4acf-a34a-3588a8bd1176\" (UID: \"260e7406-3b55-4acf-a34a-3588a8bd1176\") " Dec 09 16:12:00 crc kubenswrapper[4894]: I1209 16:12:00.928023 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/260e7406-3b55-4acf-a34a-3588a8bd1176-ceph" (OuterVolumeSpecName: "ceph") pod "260e7406-3b55-4acf-a34a-3588a8bd1176" (UID: "260e7406-3b55-4acf-a34a-3588a8bd1176"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:12:00 crc kubenswrapper[4894]: I1209 16:12:00.928261 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/260e7406-3b55-4acf-a34a-3588a8bd1176-kube-api-access-zzhxx" (OuterVolumeSpecName: "kube-api-access-zzhxx") pod "260e7406-3b55-4acf-a34a-3588a8bd1176" (UID: "260e7406-3b55-4acf-a34a-3588a8bd1176"). InnerVolumeSpecName "kube-api-access-zzhxx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:12:00 crc kubenswrapper[4894]: I1209 16:12:00.942201 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/260e7406-3b55-4acf-a34a-3588a8bd1176-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "260e7406-3b55-4acf-a34a-3588a8bd1176" (UID: "260e7406-3b55-4acf-a34a-3588a8bd1176"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:12:00 crc kubenswrapper[4894]: I1209 16:12:00.951092 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/260e7406-3b55-4acf-a34a-3588a8bd1176-inventory" (OuterVolumeSpecName: "inventory") pod "260e7406-3b55-4acf-a34a-3588a8bd1176" (UID: "260e7406-3b55-4acf-a34a-3588a8bd1176"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.016888 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzhxx\" (UniqueName: \"kubernetes.io/projected/260e7406-3b55-4acf-a34a-3588a8bd1176-kube-api-access-zzhxx\") on node \"crc\" DevicePath \"\"" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.016921 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/260e7406-3b55-4acf-a34a-3588a8bd1176-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.016931 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/260e7406-3b55-4acf-a34a-3588a8bd1176-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.016969 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/260e7406-3b55-4acf-a34a-3588a8bd1176-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.390752 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g" event={"ID":"260e7406-3b55-4acf-a34a-3588a8bd1176","Type":"ContainerDied","Data":"1723afa30179ead6279b982636b57204e333b78ee8da000e7544800e62c212d8"} Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.391005 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1723afa30179ead6279b982636b57204e333b78ee8da000e7544800e62c212d8" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.391037 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.465574 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5"] Dec 09 16:12:01 crc kubenswrapper[4894]: E1209 16:12:01.466623 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="260e7406-3b55-4acf-a34a-3588a8bd1176" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.466772 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="260e7406-3b55-4acf-a34a-3588a8bd1176" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.467097 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="260e7406-3b55-4acf-a34a-3588a8bd1176" containerName="ceph-client-edpm-deployment-openstack-edpm-ipam" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.467999 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.470406 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.470487 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-config" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.470565 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.470697 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.472031 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.472969 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.481333 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5"] Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.526032 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5b5q5\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.526114 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5b5q5\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.526168 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5b5q5\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.526198 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5b5q5\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.526258 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vhqs\" (UniqueName: \"kubernetes.io/projected/9c9358fd-4505-41a1-8c05-82c59e737fa2-kube-api-access-6vhqs\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5b5q5\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.526326 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9c9358fd-4505-41a1-8c05-82c59e737fa2-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5b5q5\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.628162 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5b5q5\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.628221 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5b5q5\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.628264 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5b5q5\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.628315 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vhqs\" (UniqueName: \"kubernetes.io/projected/9c9358fd-4505-41a1-8c05-82c59e737fa2-kube-api-access-6vhqs\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5b5q5\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.628391 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9c9358fd-4505-41a1-8c05-82c59e737fa2-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5b5q5\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.628496 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5b5q5\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.631070 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9c9358fd-4505-41a1-8c05-82c59e737fa2-ovncontroller-config-0\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5b5q5\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.633352 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-inventory\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5b5q5\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.634095 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-ovn-combined-ca-bundle\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5b5q5\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.636321 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-ssh-key\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5b5q5\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.639099 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-ceph\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5b5q5\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.653922 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vhqs\" (UniqueName: \"kubernetes.io/projected/9c9358fd-4505-41a1-8c05-82c59e737fa2-kube-api-access-6vhqs\") pod \"ovn-edpm-deployment-openstack-edpm-ipam-5b5q5\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" Dec 09 16:12:01 crc kubenswrapper[4894]: I1209 16:12:01.789735 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" Dec 09 16:12:02 crc kubenswrapper[4894]: I1209 16:12:02.277450 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5"] Dec 09 16:12:02 crc kubenswrapper[4894]: I1209 16:12:02.402380 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" event={"ID":"9c9358fd-4505-41a1-8c05-82c59e737fa2","Type":"ContainerStarted","Data":"6902bde2227f48f86452f5eff32507fc543eb2d4af3be8f09a28a90f09cb2f7c"} Dec 09 16:12:04 crc kubenswrapper[4894]: I1209 16:12:04.446354 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" event={"ID":"9c9358fd-4505-41a1-8c05-82c59e737fa2","Type":"ContainerStarted","Data":"e8dd6ffcb478aa4c7b552f53bf2bfeba36d8c1cad16447ff22643aa4fecd4f4f"} Dec 09 16:12:04 crc kubenswrapper[4894]: I1209 16:12:04.470793 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" podStartSLOduration=1.977463691 podStartE2EDuration="3.470774692s" podCreationTimestamp="2025-12-09 16:12:01 +0000 UTC" firstStartedPulling="2025-12-09 16:12:02.283384439 +0000 UTC m=+2416.602595108" lastFinishedPulling="2025-12-09 16:12:03.77669544 +0000 UTC m=+2418.095906109" observedRunningTime="2025-12-09 16:12:04.463501125 +0000 UTC m=+2418.782711834" watchObservedRunningTime="2025-12-09 16:12:04.470774692 +0000 UTC m=+2418.789985361" Dec 09 16:12:11 crc kubenswrapper[4894]: I1209 16:12:11.107123 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:12:11 crc kubenswrapper[4894]: E1209 16:12:11.108533 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:12:25 crc kubenswrapper[4894]: I1209 16:12:25.106786 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:12:25 crc kubenswrapper[4894]: E1209 16:12:25.107612 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:12:40 crc kubenswrapper[4894]: I1209 16:12:40.107176 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:12:40 crc kubenswrapper[4894]: E1209 16:12:40.108301 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:12:51 crc kubenswrapper[4894]: I1209 16:12:51.106172 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:12:51 crc kubenswrapper[4894]: E1209 16:12:51.106899 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:13:02 crc kubenswrapper[4894]: I1209 16:13:02.106303 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:13:02 crc kubenswrapper[4894]: E1209 16:13:02.107181 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:13:14 crc kubenswrapper[4894]: I1209 16:13:14.103996 4894 generic.go:334] "Generic (PLEG): container finished" podID="9c9358fd-4505-41a1-8c05-82c59e737fa2" containerID="e8dd6ffcb478aa4c7b552f53bf2bfeba36d8c1cad16447ff22643aa4fecd4f4f" exitCode=0 Dec 09 16:13:14 crc kubenswrapper[4894]: I1209 16:13:14.104070 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" event={"ID":"9c9358fd-4505-41a1-8c05-82c59e737fa2","Type":"ContainerDied","Data":"e8dd6ffcb478aa4c7b552f53bf2bfeba36d8c1cad16447ff22643aa4fecd4f4f"} Dec 09 16:13:14 crc kubenswrapper[4894]: I1209 16:13:14.107221 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:13:14 crc kubenswrapper[4894]: E1209 16:13:14.107520 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:13:15 crc kubenswrapper[4894]: I1209 16:13:15.513523 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" Dec 09 16:13:15 crc kubenswrapper[4894]: I1209 16:13:15.578977 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vhqs\" (UniqueName: \"kubernetes.io/projected/9c9358fd-4505-41a1-8c05-82c59e737fa2-kube-api-access-6vhqs\") pod \"9c9358fd-4505-41a1-8c05-82c59e737fa2\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " Dec 09 16:13:15 crc kubenswrapper[4894]: I1209 16:13:15.579053 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-ssh-key\") pod \"9c9358fd-4505-41a1-8c05-82c59e737fa2\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " Dec 09 16:13:15 crc kubenswrapper[4894]: I1209 16:13:15.579335 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-ceph\") pod \"9c9358fd-4505-41a1-8c05-82c59e737fa2\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " Dec 09 16:13:15 crc kubenswrapper[4894]: I1209 16:13:15.579399 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-inventory\") pod \"9c9358fd-4505-41a1-8c05-82c59e737fa2\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " Dec 09 16:13:15 crc kubenswrapper[4894]: I1209 16:13:15.579461 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9c9358fd-4505-41a1-8c05-82c59e737fa2-ovncontroller-config-0\") pod \"9c9358fd-4505-41a1-8c05-82c59e737fa2\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " Dec 09 16:13:15 crc kubenswrapper[4894]: I1209 16:13:15.579532 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-ovn-combined-ca-bundle\") pod \"9c9358fd-4505-41a1-8c05-82c59e737fa2\" (UID: \"9c9358fd-4505-41a1-8c05-82c59e737fa2\") " Dec 09 16:13:15 crc kubenswrapper[4894]: I1209 16:13:15.585286 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c9358fd-4505-41a1-8c05-82c59e737fa2-kube-api-access-6vhqs" (OuterVolumeSpecName: "kube-api-access-6vhqs") pod "9c9358fd-4505-41a1-8c05-82c59e737fa2" (UID: "9c9358fd-4505-41a1-8c05-82c59e737fa2"). InnerVolumeSpecName "kube-api-access-6vhqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:13:15 crc kubenswrapper[4894]: I1209 16:13:15.585662 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-ovn-combined-ca-bundle" (OuterVolumeSpecName: "ovn-combined-ca-bundle") pod "9c9358fd-4505-41a1-8c05-82c59e737fa2" (UID: "9c9358fd-4505-41a1-8c05-82c59e737fa2"). InnerVolumeSpecName "ovn-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:13:15 crc kubenswrapper[4894]: I1209 16:13:15.599755 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-ceph" (OuterVolumeSpecName: "ceph") pod "9c9358fd-4505-41a1-8c05-82c59e737fa2" (UID: "9c9358fd-4505-41a1-8c05-82c59e737fa2"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:13:15 crc kubenswrapper[4894]: I1209 16:13:15.608045 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-inventory" (OuterVolumeSpecName: "inventory") pod "9c9358fd-4505-41a1-8c05-82c59e737fa2" (UID: "9c9358fd-4505-41a1-8c05-82c59e737fa2"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:13:15 crc kubenswrapper[4894]: I1209 16:13:15.611981 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9c9358fd-4505-41a1-8c05-82c59e737fa2-ovncontroller-config-0" (OuterVolumeSpecName: "ovncontroller-config-0") pod "9c9358fd-4505-41a1-8c05-82c59e737fa2" (UID: "9c9358fd-4505-41a1-8c05-82c59e737fa2"). InnerVolumeSpecName "ovncontroller-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:13:15 crc kubenswrapper[4894]: I1209 16:13:15.613196 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9c9358fd-4505-41a1-8c05-82c59e737fa2" (UID: "9c9358fd-4505-41a1-8c05-82c59e737fa2"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:13:15 crc kubenswrapper[4894]: I1209 16:13:15.680953 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:13:15 crc kubenswrapper[4894]: I1209 16:13:15.680989 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 16:13:15 crc kubenswrapper[4894]: I1209 16:13:15.681000 4894 reconciler_common.go:293] "Volume detached for volume \"ovncontroller-config-0\" (UniqueName: \"kubernetes.io/configmap/9c9358fd-4505-41a1-8c05-82c59e737fa2-ovncontroller-config-0\") on node \"crc\" DevicePath \"\"" Dec 09 16:13:15 crc kubenswrapper[4894]: I1209 16:13:15.681009 4894 reconciler_common.go:293] "Volume detached for volume \"ovn-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-ovn-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:13:15 crc kubenswrapper[4894]: I1209 16:13:15.681018 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vhqs\" (UniqueName: \"kubernetes.io/projected/9c9358fd-4505-41a1-8c05-82c59e737fa2-kube-api-access-6vhqs\") on node \"crc\" DevicePath \"\"" Dec 09 16:13:15 crc kubenswrapper[4894]: I1209 16:13:15.681026 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9c9358fd-4505-41a1-8c05-82c59e737fa2-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.121255 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" event={"ID":"9c9358fd-4505-41a1-8c05-82c59e737fa2","Type":"ContainerDied","Data":"6902bde2227f48f86452f5eff32507fc543eb2d4af3be8f09a28a90f09cb2f7c"} Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.121785 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6902bde2227f48f86452f5eff32507fc543eb2d4af3be8f09a28a90f09cb2f7c" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.121308 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-edpm-deployment-openstack-edpm-ipam-5b5q5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.229223 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5"] Dec 09 16:13:16 crc kubenswrapper[4894]: E1209 16:13:16.230091 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c9358fd-4505-41a1-8c05-82c59e737fa2" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.230143 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c9358fd-4505-41a1-8c05-82c59e737fa2" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.230925 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c9358fd-4505-41a1-8c05-82c59e737fa2" containerName="ovn-edpm-deployment-openstack-edpm-ipam" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.232177 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.235132 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-neutron-config" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.235733 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.235913 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.236121 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.236606 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.237235 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-ovn-metadata-agent-neutron-config" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.237545 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.252471 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5"] Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.292999 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.293066 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.293093 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-prh26\" (UniqueName: \"kubernetes.io/projected/be6e221e-e719-4805-9198-0bc9862fa4d0-kube-api-access-prh26\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.293136 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.293155 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.293335 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.293411 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.394854 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.394919 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.394976 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.395021 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.395046 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-prh26\" (UniqueName: \"kubernetes.io/projected/be6e221e-e719-4805-9198-0bc9862fa4d0-kube-api-access-prh26\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.395083 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.395105 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.400544 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-nova-metadata-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.401201 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-inventory\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.401850 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-ssh-key\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.404094 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-ceph\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.408288 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-neutron-metadata-combined-ca-bundle\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.410401 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-neutron-ovn-metadata-agent-neutron-config-0\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.415602 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-prh26\" (UniqueName: \"kubernetes.io/projected/be6e221e-e719-4805-9198-0bc9862fa4d0-kube-api-access-prh26\") pod \"neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:16 crc kubenswrapper[4894]: I1209 16:13:16.571888 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:13:17 crc kubenswrapper[4894]: I1209 16:13:17.117501 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5"] Dec 09 16:13:18 crc kubenswrapper[4894]: I1209 16:13:18.146550 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" event={"ID":"be6e221e-e719-4805-9198-0bc9862fa4d0","Type":"ContainerStarted","Data":"838f403930436d0402d32e2eda6309d4af18161508f5ba25f276b9002ae7594a"} Dec 09 16:13:18 crc kubenswrapper[4894]: I1209 16:13:18.147774 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" event={"ID":"be6e221e-e719-4805-9198-0bc9862fa4d0","Type":"ContainerStarted","Data":"81b450742d2d2da1685847a266bca2cd455abf5381b679a52ab4f8c2a41ed5c0"} Dec 09 16:13:18 crc kubenswrapper[4894]: I1209 16:13:18.172995 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" podStartSLOduration=1.492049283 podStartE2EDuration="2.172964428s" podCreationTimestamp="2025-12-09 16:13:16 +0000 UTC" firstStartedPulling="2025-12-09 16:13:17.128105735 +0000 UTC m=+2491.447316404" lastFinishedPulling="2025-12-09 16:13:17.80902085 +0000 UTC m=+2492.128231549" observedRunningTime="2025-12-09 16:13:18.171026986 +0000 UTC m=+2492.490237665" watchObservedRunningTime="2025-12-09 16:13:18.172964428 +0000 UTC m=+2492.492175127" Dec 09 16:13:29 crc kubenswrapper[4894]: I1209 16:13:29.106300 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:13:29 crc kubenswrapper[4894]: E1209 16:13:29.107121 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:13:43 crc kubenswrapper[4894]: I1209 16:13:43.107371 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:13:43 crc kubenswrapper[4894]: I1209 16:13:43.382963 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerStarted","Data":"528a0efd27e59e2fb3d60ccf4dc2db0b0f33df50d99afe9391f928ded08456ba"} Dec 09 16:13:48 crc kubenswrapper[4894]: I1209 16:13:48.003670 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-vbx5k"] Dec 09 16:13:48 crc kubenswrapper[4894]: I1209 16:13:48.006469 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vbx5k" Dec 09 16:13:48 crc kubenswrapper[4894]: I1209 16:13:48.033450 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vbx5k"] Dec 09 16:13:48 crc kubenswrapper[4894]: I1209 16:13:48.147505 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3040ccec-a87e-401d-bd24-1449cb73cbe5-catalog-content\") pod \"certified-operators-vbx5k\" (UID: \"3040ccec-a87e-401d-bd24-1449cb73cbe5\") " pod="openshift-marketplace/certified-operators-vbx5k" Dec 09 16:13:48 crc kubenswrapper[4894]: I1209 16:13:48.147766 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3040ccec-a87e-401d-bd24-1449cb73cbe5-utilities\") pod \"certified-operators-vbx5k\" (UID: \"3040ccec-a87e-401d-bd24-1449cb73cbe5\") " pod="openshift-marketplace/certified-operators-vbx5k" Dec 09 16:13:48 crc kubenswrapper[4894]: I1209 16:13:48.147996 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8ns5\" (UniqueName: \"kubernetes.io/projected/3040ccec-a87e-401d-bd24-1449cb73cbe5-kube-api-access-h8ns5\") pod \"certified-operators-vbx5k\" (UID: \"3040ccec-a87e-401d-bd24-1449cb73cbe5\") " pod="openshift-marketplace/certified-operators-vbx5k" Dec 09 16:13:48 crc kubenswrapper[4894]: I1209 16:13:48.250492 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3040ccec-a87e-401d-bd24-1449cb73cbe5-catalog-content\") pod \"certified-operators-vbx5k\" (UID: \"3040ccec-a87e-401d-bd24-1449cb73cbe5\") " pod="openshift-marketplace/certified-operators-vbx5k" Dec 09 16:13:48 crc kubenswrapper[4894]: I1209 16:13:48.250815 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3040ccec-a87e-401d-bd24-1449cb73cbe5-utilities\") pod \"certified-operators-vbx5k\" (UID: \"3040ccec-a87e-401d-bd24-1449cb73cbe5\") " pod="openshift-marketplace/certified-operators-vbx5k" Dec 09 16:13:48 crc kubenswrapper[4894]: I1209 16:13:48.250892 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8ns5\" (UniqueName: \"kubernetes.io/projected/3040ccec-a87e-401d-bd24-1449cb73cbe5-kube-api-access-h8ns5\") pod \"certified-operators-vbx5k\" (UID: \"3040ccec-a87e-401d-bd24-1449cb73cbe5\") " pod="openshift-marketplace/certified-operators-vbx5k" Dec 09 16:13:48 crc kubenswrapper[4894]: I1209 16:13:48.251057 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3040ccec-a87e-401d-bd24-1449cb73cbe5-catalog-content\") pod \"certified-operators-vbx5k\" (UID: \"3040ccec-a87e-401d-bd24-1449cb73cbe5\") " pod="openshift-marketplace/certified-operators-vbx5k" Dec 09 16:13:48 crc kubenswrapper[4894]: I1209 16:13:48.251775 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3040ccec-a87e-401d-bd24-1449cb73cbe5-utilities\") pod \"certified-operators-vbx5k\" (UID: \"3040ccec-a87e-401d-bd24-1449cb73cbe5\") " pod="openshift-marketplace/certified-operators-vbx5k" Dec 09 16:13:48 crc kubenswrapper[4894]: I1209 16:13:48.274983 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8ns5\" (UniqueName: \"kubernetes.io/projected/3040ccec-a87e-401d-bd24-1449cb73cbe5-kube-api-access-h8ns5\") pod \"certified-operators-vbx5k\" (UID: \"3040ccec-a87e-401d-bd24-1449cb73cbe5\") " pod="openshift-marketplace/certified-operators-vbx5k" Dec 09 16:13:48 crc kubenswrapper[4894]: I1209 16:13:48.341736 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vbx5k" Dec 09 16:13:48 crc kubenswrapper[4894]: I1209 16:13:48.875196 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-vbx5k"] Dec 09 16:13:48 crc kubenswrapper[4894]: W1209 16:13:48.885762 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3040ccec_a87e_401d_bd24_1449cb73cbe5.slice/crio-83df7784571eb34a763432215663f5351687e4f63d48732c1f7f4bfa1cd6ac61 WatchSource:0}: Error finding container 83df7784571eb34a763432215663f5351687e4f63d48732c1f7f4bfa1cd6ac61: Status 404 returned error can't find the container with id 83df7784571eb34a763432215663f5351687e4f63d48732c1f7f4bfa1cd6ac61 Dec 09 16:13:49 crc kubenswrapper[4894]: I1209 16:13:49.441669 4894 generic.go:334] "Generic (PLEG): container finished" podID="3040ccec-a87e-401d-bd24-1449cb73cbe5" containerID="9c2bb7b3adfe3e1379ee285dd7300ffebfddb72b3755e1424ddb2a23314ceef7" exitCode=0 Dec 09 16:13:49 crc kubenswrapper[4894]: I1209 16:13:49.441779 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbx5k" event={"ID":"3040ccec-a87e-401d-bd24-1449cb73cbe5","Type":"ContainerDied","Data":"9c2bb7b3adfe3e1379ee285dd7300ffebfddb72b3755e1424ddb2a23314ceef7"} Dec 09 16:13:49 crc kubenswrapper[4894]: I1209 16:13:49.442814 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbx5k" event={"ID":"3040ccec-a87e-401d-bd24-1449cb73cbe5","Type":"ContainerStarted","Data":"83df7784571eb34a763432215663f5351687e4f63d48732c1f7f4bfa1cd6ac61"} Dec 09 16:13:49 crc kubenswrapper[4894]: I1209 16:13:49.444220 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 16:13:50 crc kubenswrapper[4894]: I1209 16:13:50.457206 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbx5k" event={"ID":"3040ccec-a87e-401d-bd24-1449cb73cbe5","Type":"ContainerStarted","Data":"8df571227ab7b0c741c694b6cb2ae85eb0bac27d184f39e5d7de6c1575eb2b2b"} Dec 09 16:13:51 crc kubenswrapper[4894]: I1209 16:13:51.474363 4894 generic.go:334] "Generic (PLEG): container finished" podID="3040ccec-a87e-401d-bd24-1449cb73cbe5" containerID="8df571227ab7b0c741c694b6cb2ae85eb0bac27d184f39e5d7de6c1575eb2b2b" exitCode=0 Dec 09 16:13:51 crc kubenswrapper[4894]: I1209 16:13:51.474414 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbx5k" event={"ID":"3040ccec-a87e-401d-bd24-1449cb73cbe5","Type":"ContainerDied","Data":"8df571227ab7b0c741c694b6cb2ae85eb0bac27d184f39e5d7de6c1575eb2b2b"} Dec 09 16:13:52 crc kubenswrapper[4894]: I1209 16:13:52.486715 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbx5k" event={"ID":"3040ccec-a87e-401d-bd24-1449cb73cbe5","Type":"ContainerStarted","Data":"25d65d48d042fbcb8bf8cc42e31a8d22f2c140b425f9bc9337db00ae456f9fb4"} Dec 09 16:13:58 crc kubenswrapper[4894]: I1209 16:13:58.342388 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-vbx5k" Dec 09 16:13:58 crc kubenswrapper[4894]: I1209 16:13:58.342913 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-vbx5k" Dec 09 16:13:58 crc kubenswrapper[4894]: I1209 16:13:58.399587 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-vbx5k" Dec 09 16:13:58 crc kubenswrapper[4894]: I1209 16:13:58.435531 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-vbx5k" podStartSLOduration=8.744880083 podStartE2EDuration="11.435502967s" podCreationTimestamp="2025-12-09 16:13:47 +0000 UTC" firstStartedPulling="2025-12-09 16:13:49.44384719 +0000 UTC m=+2523.763057889" lastFinishedPulling="2025-12-09 16:13:52.134470094 +0000 UTC m=+2526.453680773" observedRunningTime="2025-12-09 16:13:52.511858577 +0000 UTC m=+2526.831069286" watchObservedRunningTime="2025-12-09 16:13:58.435502967 +0000 UTC m=+2532.754713676" Dec 09 16:13:58 crc kubenswrapper[4894]: I1209 16:13:58.629542 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-vbx5k" Dec 09 16:13:58 crc kubenswrapper[4894]: I1209 16:13:58.697891 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vbx5k"] Dec 09 16:14:00 crc kubenswrapper[4894]: I1209 16:14:00.575328 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-vbx5k" podUID="3040ccec-a87e-401d-bd24-1449cb73cbe5" containerName="registry-server" containerID="cri-o://25d65d48d042fbcb8bf8cc42e31a8d22f2c140b425f9bc9337db00ae456f9fb4" gracePeriod=2 Dec 09 16:14:03 crc kubenswrapper[4894]: I1209 16:14:03.617505 4894 generic.go:334] "Generic (PLEG): container finished" podID="3040ccec-a87e-401d-bd24-1449cb73cbe5" containerID="25d65d48d042fbcb8bf8cc42e31a8d22f2c140b425f9bc9337db00ae456f9fb4" exitCode=0 Dec 09 16:14:03 crc kubenswrapper[4894]: I1209 16:14:03.617692 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbx5k" event={"ID":"3040ccec-a87e-401d-bd24-1449cb73cbe5","Type":"ContainerDied","Data":"25d65d48d042fbcb8bf8cc42e31a8d22f2c140b425f9bc9337db00ae456f9fb4"} Dec 09 16:14:03 crc kubenswrapper[4894]: I1209 16:14:03.772502 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vbx5k" Dec 09 16:14:03 crc kubenswrapper[4894]: I1209 16:14:03.932359 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h8ns5\" (UniqueName: \"kubernetes.io/projected/3040ccec-a87e-401d-bd24-1449cb73cbe5-kube-api-access-h8ns5\") pod \"3040ccec-a87e-401d-bd24-1449cb73cbe5\" (UID: \"3040ccec-a87e-401d-bd24-1449cb73cbe5\") " Dec 09 16:14:03 crc kubenswrapper[4894]: I1209 16:14:03.932505 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3040ccec-a87e-401d-bd24-1449cb73cbe5-utilities\") pod \"3040ccec-a87e-401d-bd24-1449cb73cbe5\" (UID: \"3040ccec-a87e-401d-bd24-1449cb73cbe5\") " Dec 09 16:14:03 crc kubenswrapper[4894]: I1209 16:14:03.932595 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3040ccec-a87e-401d-bd24-1449cb73cbe5-catalog-content\") pod \"3040ccec-a87e-401d-bd24-1449cb73cbe5\" (UID: \"3040ccec-a87e-401d-bd24-1449cb73cbe5\") " Dec 09 16:14:03 crc kubenswrapper[4894]: I1209 16:14:03.934102 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3040ccec-a87e-401d-bd24-1449cb73cbe5-utilities" (OuterVolumeSpecName: "utilities") pod "3040ccec-a87e-401d-bd24-1449cb73cbe5" (UID: "3040ccec-a87e-401d-bd24-1449cb73cbe5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:14:03 crc kubenswrapper[4894]: I1209 16:14:03.952910 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3040ccec-a87e-401d-bd24-1449cb73cbe5-kube-api-access-h8ns5" (OuterVolumeSpecName: "kube-api-access-h8ns5") pod "3040ccec-a87e-401d-bd24-1449cb73cbe5" (UID: "3040ccec-a87e-401d-bd24-1449cb73cbe5"). InnerVolumeSpecName "kube-api-access-h8ns5". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:14:04 crc kubenswrapper[4894]: I1209 16:14:04.007040 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3040ccec-a87e-401d-bd24-1449cb73cbe5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3040ccec-a87e-401d-bd24-1449cb73cbe5" (UID: "3040ccec-a87e-401d-bd24-1449cb73cbe5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:14:04 crc kubenswrapper[4894]: I1209 16:14:04.034541 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h8ns5\" (UniqueName: \"kubernetes.io/projected/3040ccec-a87e-401d-bd24-1449cb73cbe5-kube-api-access-h8ns5\") on node \"crc\" DevicePath \"\"" Dec 09 16:14:04 crc kubenswrapper[4894]: I1209 16:14:04.034582 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3040ccec-a87e-401d-bd24-1449cb73cbe5-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:14:04 crc kubenswrapper[4894]: I1209 16:14:04.034595 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3040ccec-a87e-401d-bd24-1449cb73cbe5-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:14:04 crc kubenswrapper[4894]: I1209 16:14:04.639005 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-vbx5k" event={"ID":"3040ccec-a87e-401d-bd24-1449cb73cbe5","Type":"ContainerDied","Data":"83df7784571eb34a763432215663f5351687e4f63d48732c1f7f4bfa1cd6ac61"} Dec 09 16:14:04 crc kubenswrapper[4894]: I1209 16:14:04.639159 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-vbx5k" Dec 09 16:14:04 crc kubenswrapper[4894]: I1209 16:14:04.639687 4894 scope.go:117] "RemoveContainer" containerID="25d65d48d042fbcb8bf8cc42e31a8d22f2c140b425f9bc9337db00ae456f9fb4" Dec 09 16:14:04 crc kubenswrapper[4894]: I1209 16:14:04.675792 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-vbx5k"] Dec 09 16:14:04 crc kubenswrapper[4894]: I1209 16:14:04.687616 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-vbx5k"] Dec 09 16:14:04 crc kubenswrapper[4894]: I1209 16:14:04.698243 4894 scope.go:117] "RemoveContainer" containerID="8df571227ab7b0c741c694b6cb2ae85eb0bac27d184f39e5d7de6c1575eb2b2b" Dec 09 16:14:04 crc kubenswrapper[4894]: I1209 16:14:04.724160 4894 scope.go:117] "RemoveContainer" containerID="9c2bb7b3adfe3e1379ee285dd7300ffebfddb72b3755e1424ddb2a23314ceef7" Dec 09 16:14:06 crc kubenswrapper[4894]: I1209 16:14:06.128272 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3040ccec-a87e-401d-bd24-1449cb73cbe5" path="/var/lib/kubelet/pods/3040ccec-a87e-401d-bd24-1449cb73cbe5/volumes" Dec 09 16:14:16 crc kubenswrapper[4894]: I1209 16:14:16.806384 4894 generic.go:334] "Generic (PLEG): container finished" podID="be6e221e-e719-4805-9198-0bc9862fa4d0" containerID="838f403930436d0402d32e2eda6309d4af18161508f5ba25f276b9002ae7594a" exitCode=0 Dec 09 16:14:16 crc kubenswrapper[4894]: I1209 16:14:16.806665 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" event={"ID":"be6e221e-e719-4805-9198-0bc9862fa4d0","Type":"ContainerDied","Data":"838f403930436d0402d32e2eda6309d4af18161508f5ba25f276b9002ae7594a"} Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.242360 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.329324 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-ceph\") pod \"be6e221e-e719-4805-9198-0bc9862fa4d0\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.329379 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-neutron-ovn-metadata-agent-neutron-config-0\") pod \"be6e221e-e719-4805-9198-0bc9862fa4d0\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.329439 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-nova-metadata-neutron-config-0\") pod \"be6e221e-e719-4805-9198-0bc9862fa4d0\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.329461 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-prh26\" (UniqueName: \"kubernetes.io/projected/be6e221e-e719-4805-9198-0bc9862fa4d0-kube-api-access-prh26\") pod \"be6e221e-e719-4805-9198-0bc9862fa4d0\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.329483 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-neutron-metadata-combined-ca-bundle\") pod \"be6e221e-e719-4805-9198-0bc9862fa4d0\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.329521 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-inventory\") pod \"be6e221e-e719-4805-9198-0bc9862fa4d0\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.329568 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-ssh-key\") pod \"be6e221e-e719-4805-9198-0bc9862fa4d0\" (UID: \"be6e221e-e719-4805-9198-0bc9862fa4d0\") " Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.336170 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-neutron-metadata-combined-ca-bundle" (OuterVolumeSpecName: "neutron-metadata-combined-ca-bundle") pod "be6e221e-e719-4805-9198-0bc9862fa4d0" (UID: "be6e221e-e719-4805-9198-0bc9862fa4d0"). InnerVolumeSpecName "neutron-metadata-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.344601 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be6e221e-e719-4805-9198-0bc9862fa4d0-kube-api-access-prh26" (OuterVolumeSpecName: "kube-api-access-prh26") pod "be6e221e-e719-4805-9198-0bc9862fa4d0" (UID: "be6e221e-e719-4805-9198-0bc9862fa4d0"). InnerVolumeSpecName "kube-api-access-prh26". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.351964 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-ceph" (OuterVolumeSpecName: "ceph") pod "be6e221e-e719-4805-9198-0bc9862fa4d0" (UID: "be6e221e-e719-4805-9198-0bc9862fa4d0"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.374981 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-neutron-ovn-metadata-agent-neutron-config-0" (OuterVolumeSpecName: "neutron-ovn-metadata-agent-neutron-config-0") pod "be6e221e-e719-4805-9198-0bc9862fa4d0" (UID: "be6e221e-e719-4805-9198-0bc9862fa4d0"). InnerVolumeSpecName "neutron-ovn-metadata-agent-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.375945 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-inventory" (OuterVolumeSpecName: "inventory") pod "be6e221e-e719-4805-9198-0bc9862fa4d0" (UID: "be6e221e-e719-4805-9198-0bc9862fa4d0"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.382291 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-nova-metadata-neutron-config-0" (OuterVolumeSpecName: "nova-metadata-neutron-config-0") pod "be6e221e-e719-4805-9198-0bc9862fa4d0" (UID: "be6e221e-e719-4805-9198-0bc9862fa4d0"). InnerVolumeSpecName "nova-metadata-neutron-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.384886 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "be6e221e-e719-4805-9198-0bc9862fa4d0" (UID: "be6e221e-e719-4805-9198-0bc9862fa4d0"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.432092 4894 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-nova-metadata-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.432146 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-prh26\" (UniqueName: \"kubernetes.io/projected/be6e221e-e719-4805-9198-0bc9862fa4d0-kube-api-access-prh26\") on node \"crc\" DevicePath \"\"" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.432168 4894 reconciler_common.go:293] "Volume detached for volume \"neutron-metadata-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-neutron-metadata-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.432190 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.432207 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.432226 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.432246 4894 reconciler_common.go:293] "Volume detached for volume \"neutron-ovn-metadata-agent-neutron-config-0\" (UniqueName: \"kubernetes.io/secret/be6e221e-e719-4805-9198-0bc9862fa4d0-neutron-ovn-metadata-agent-neutron-config-0\") on node \"crc\" DevicePath \"\"" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.834462 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" event={"ID":"be6e221e-e719-4805-9198-0bc9862fa4d0","Type":"ContainerDied","Data":"81b450742d2d2da1685847a266bca2cd455abf5381b679a52ab4f8c2a41ed5c0"} Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.834826 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="81b450742d2d2da1685847a266bca2cd455abf5381b679a52ab4f8c2a41ed5c0" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.834569 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.953230 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk"] Dec 09 16:14:18 crc kubenswrapper[4894]: E1209 16:14:18.953806 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3040ccec-a87e-401d-bd24-1449cb73cbe5" containerName="extract-content" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.953828 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="3040ccec-a87e-401d-bd24-1449cb73cbe5" containerName="extract-content" Dec 09 16:14:18 crc kubenswrapper[4894]: E1209 16:14:18.953845 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3040ccec-a87e-401d-bd24-1449cb73cbe5" containerName="registry-server" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.953853 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="3040ccec-a87e-401d-bd24-1449cb73cbe5" containerName="registry-server" Dec 09 16:14:18 crc kubenswrapper[4894]: E1209 16:14:18.953870 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3040ccec-a87e-401d-bd24-1449cb73cbe5" containerName="extract-utilities" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.953879 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="3040ccec-a87e-401d-bd24-1449cb73cbe5" containerName="extract-utilities" Dec 09 16:14:18 crc kubenswrapper[4894]: E1209 16:14:18.953890 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be6e221e-e719-4805-9198-0bc9862fa4d0" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.953900 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="be6e221e-e719-4805-9198-0bc9862fa4d0" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.954105 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="3040ccec-a87e-401d-bd24-1449cb73cbe5" containerName="registry-server" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.954143 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="be6e221e-e719-4805-9198-0bc9862fa4d0" containerName="neutron-metadata-edpm-deployment-openstack-edpm-ipam" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.954872 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.956544 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.958599 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.958693 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"libvirt-secret" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.961804 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.962436 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.962802 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 16:14:18 crc kubenswrapper[4894]: I1209 16:14:18.967913 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk"] Dec 09 16:14:19 crc kubenswrapper[4894]: I1209 16:14:19.144212 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjlv8\" (UniqueName: \"kubernetes.io/projected/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-kube-api-access-sjlv8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" Dec 09 16:14:19 crc kubenswrapper[4894]: I1209 16:14:19.144324 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" Dec 09 16:14:19 crc kubenswrapper[4894]: I1209 16:14:19.144878 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" Dec 09 16:14:19 crc kubenswrapper[4894]: I1209 16:14:19.145004 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" Dec 09 16:14:19 crc kubenswrapper[4894]: I1209 16:14:19.145053 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" Dec 09 16:14:19 crc kubenswrapper[4894]: I1209 16:14:19.145366 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" Dec 09 16:14:19 crc kubenswrapper[4894]: I1209 16:14:19.247332 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" Dec 09 16:14:19 crc kubenswrapper[4894]: I1209 16:14:19.247383 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" Dec 09 16:14:19 crc kubenswrapper[4894]: I1209 16:14:19.247431 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" Dec 09 16:14:19 crc kubenswrapper[4894]: I1209 16:14:19.247532 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjlv8\" (UniqueName: \"kubernetes.io/projected/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-kube-api-access-sjlv8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" Dec 09 16:14:19 crc kubenswrapper[4894]: I1209 16:14:19.247565 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" Dec 09 16:14:19 crc kubenswrapper[4894]: I1209 16:14:19.247616 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" Dec 09 16:14:19 crc kubenswrapper[4894]: I1209 16:14:19.255390 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-libvirt-secret-0\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" Dec 09 16:14:19 crc kubenswrapper[4894]: I1209 16:14:19.255633 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-inventory\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" Dec 09 16:14:19 crc kubenswrapper[4894]: I1209 16:14:19.255991 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-libvirt-combined-ca-bundle\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" Dec 09 16:14:19 crc kubenswrapper[4894]: I1209 16:14:19.261329 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-ssh-key\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" Dec 09 16:14:19 crc kubenswrapper[4894]: I1209 16:14:19.261830 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-ceph\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" Dec 09 16:14:19 crc kubenswrapper[4894]: I1209 16:14:19.286559 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjlv8\" (UniqueName: \"kubernetes.io/projected/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-kube-api-access-sjlv8\") pod \"libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" Dec 09 16:14:19 crc kubenswrapper[4894]: I1209 16:14:19.291791 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" Dec 09 16:14:19 crc kubenswrapper[4894]: I1209 16:14:19.912668 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk"] Dec 09 16:14:20 crc kubenswrapper[4894]: I1209 16:14:20.860584 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" event={"ID":"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1","Type":"ContainerStarted","Data":"c0ba307bd08f307f88d2d21f9ea10c64a25718a2a0f1cd76d5882d7936b1d24e"} Dec 09 16:14:21 crc kubenswrapper[4894]: I1209 16:14:21.874144 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" event={"ID":"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1","Type":"ContainerStarted","Data":"7cac2f4617590ede70e797c27bfc2383a299d4c01c1a66edba1153d65296f729"} Dec 09 16:14:21 crc kubenswrapper[4894]: I1209 16:14:21.914204 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" podStartSLOduration=2.735355831 podStartE2EDuration="3.914180403s" podCreationTimestamp="2025-12-09 16:14:18 +0000 UTC" firstStartedPulling="2025-12-09 16:14:19.923623512 +0000 UTC m=+2554.242834191" lastFinishedPulling="2025-12-09 16:14:21.102448044 +0000 UTC m=+2555.421658763" observedRunningTime="2025-12-09 16:14:21.899657609 +0000 UTC m=+2556.218868308" watchObservedRunningTime="2025-12-09 16:14:21.914180403 +0000 UTC m=+2556.233391112" Dec 09 16:15:00 crc kubenswrapper[4894]: I1209 16:15:00.171677 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421615-fdjmw"] Dec 09 16:15:00 crc kubenswrapper[4894]: I1209 16:15:00.173493 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421615-fdjmw" Dec 09 16:15:00 crc kubenswrapper[4894]: I1209 16:15:00.175547 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 16:15:00 crc kubenswrapper[4894]: I1209 16:15:00.187407 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 16:15:00 crc kubenswrapper[4894]: I1209 16:15:00.187898 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421615-fdjmw"] Dec 09 16:15:00 crc kubenswrapper[4894]: I1209 16:15:00.188089 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzgcx\" (UniqueName: \"kubernetes.io/projected/15be4c3e-4494-46ff-aca6-092e3eaef576-kube-api-access-hzgcx\") pod \"collect-profiles-29421615-fdjmw\" (UID: \"15be4c3e-4494-46ff-aca6-092e3eaef576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421615-fdjmw" Dec 09 16:15:00 crc kubenswrapper[4894]: I1209 16:15:00.188171 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/15be4c3e-4494-46ff-aca6-092e3eaef576-secret-volume\") pod \"collect-profiles-29421615-fdjmw\" (UID: \"15be4c3e-4494-46ff-aca6-092e3eaef576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421615-fdjmw" Dec 09 16:15:00 crc kubenswrapper[4894]: I1209 16:15:00.188274 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15be4c3e-4494-46ff-aca6-092e3eaef576-config-volume\") pod \"collect-profiles-29421615-fdjmw\" (UID: \"15be4c3e-4494-46ff-aca6-092e3eaef576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421615-fdjmw" Dec 09 16:15:00 crc kubenswrapper[4894]: I1209 16:15:00.290171 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzgcx\" (UniqueName: \"kubernetes.io/projected/15be4c3e-4494-46ff-aca6-092e3eaef576-kube-api-access-hzgcx\") pod \"collect-profiles-29421615-fdjmw\" (UID: \"15be4c3e-4494-46ff-aca6-092e3eaef576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421615-fdjmw" Dec 09 16:15:00 crc kubenswrapper[4894]: I1209 16:15:00.290270 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/15be4c3e-4494-46ff-aca6-092e3eaef576-secret-volume\") pod \"collect-profiles-29421615-fdjmw\" (UID: \"15be4c3e-4494-46ff-aca6-092e3eaef576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421615-fdjmw" Dec 09 16:15:00 crc kubenswrapper[4894]: I1209 16:15:00.290338 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15be4c3e-4494-46ff-aca6-092e3eaef576-config-volume\") pod \"collect-profiles-29421615-fdjmw\" (UID: \"15be4c3e-4494-46ff-aca6-092e3eaef576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421615-fdjmw" Dec 09 16:15:00 crc kubenswrapper[4894]: I1209 16:15:00.291269 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15be4c3e-4494-46ff-aca6-092e3eaef576-config-volume\") pod \"collect-profiles-29421615-fdjmw\" (UID: \"15be4c3e-4494-46ff-aca6-092e3eaef576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421615-fdjmw" Dec 09 16:15:00 crc kubenswrapper[4894]: I1209 16:15:00.298263 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/15be4c3e-4494-46ff-aca6-092e3eaef576-secret-volume\") pod \"collect-profiles-29421615-fdjmw\" (UID: \"15be4c3e-4494-46ff-aca6-092e3eaef576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421615-fdjmw" Dec 09 16:15:00 crc kubenswrapper[4894]: I1209 16:15:00.306946 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzgcx\" (UniqueName: \"kubernetes.io/projected/15be4c3e-4494-46ff-aca6-092e3eaef576-kube-api-access-hzgcx\") pod \"collect-profiles-29421615-fdjmw\" (UID: \"15be4c3e-4494-46ff-aca6-092e3eaef576\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421615-fdjmw" Dec 09 16:15:00 crc kubenswrapper[4894]: I1209 16:15:00.493615 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421615-fdjmw" Dec 09 16:15:00 crc kubenswrapper[4894]: I1209 16:15:00.961371 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421615-fdjmw"] Dec 09 16:15:01 crc kubenswrapper[4894]: I1209 16:15:01.270435 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421615-fdjmw" event={"ID":"15be4c3e-4494-46ff-aca6-092e3eaef576","Type":"ContainerStarted","Data":"21c0545d1886dcef4bc3a763e357e3b2614df5e4b6945fa2ecc95628609e5c8c"} Dec 09 16:15:03 crc kubenswrapper[4894]: I1209 16:15:03.555143 4894 generic.go:334] "Generic (PLEG): container finished" podID="15be4c3e-4494-46ff-aca6-092e3eaef576" containerID="b543c54301bb6bfd7f80f910af2d33a73285a37147543b001326a7f20faa4eee" exitCode=0 Dec 09 16:15:03 crc kubenswrapper[4894]: I1209 16:15:03.555188 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421615-fdjmw" event={"ID":"15be4c3e-4494-46ff-aca6-092e3eaef576","Type":"ContainerDied","Data":"b543c54301bb6bfd7f80f910af2d33a73285a37147543b001326a7f20faa4eee"} Dec 09 16:15:04 crc kubenswrapper[4894]: I1209 16:15:04.932459 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421615-fdjmw" Dec 09 16:15:05 crc kubenswrapper[4894]: I1209 16:15:05.112059 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15be4c3e-4494-46ff-aca6-092e3eaef576-config-volume\") pod \"15be4c3e-4494-46ff-aca6-092e3eaef576\" (UID: \"15be4c3e-4494-46ff-aca6-092e3eaef576\") " Dec 09 16:15:05 crc kubenswrapper[4894]: I1209 16:15:05.112795 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/15be4c3e-4494-46ff-aca6-092e3eaef576-secret-volume\") pod \"15be4c3e-4494-46ff-aca6-092e3eaef576\" (UID: \"15be4c3e-4494-46ff-aca6-092e3eaef576\") " Dec 09 16:15:05 crc kubenswrapper[4894]: I1209 16:15:05.112934 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hzgcx\" (UniqueName: \"kubernetes.io/projected/15be4c3e-4494-46ff-aca6-092e3eaef576-kube-api-access-hzgcx\") pod \"15be4c3e-4494-46ff-aca6-092e3eaef576\" (UID: \"15be4c3e-4494-46ff-aca6-092e3eaef576\") " Dec 09 16:15:05 crc kubenswrapper[4894]: I1209 16:15:05.113253 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15be4c3e-4494-46ff-aca6-092e3eaef576-config-volume" (OuterVolumeSpecName: "config-volume") pod "15be4c3e-4494-46ff-aca6-092e3eaef576" (UID: "15be4c3e-4494-46ff-aca6-092e3eaef576"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:15:05 crc kubenswrapper[4894]: I1209 16:15:05.114789 4894 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/15be4c3e-4494-46ff-aca6-092e3eaef576-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 16:15:05 crc kubenswrapper[4894]: I1209 16:15:05.122899 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15be4c3e-4494-46ff-aca6-092e3eaef576-kube-api-access-hzgcx" (OuterVolumeSpecName: "kube-api-access-hzgcx") pod "15be4c3e-4494-46ff-aca6-092e3eaef576" (UID: "15be4c3e-4494-46ff-aca6-092e3eaef576"). InnerVolumeSpecName "kube-api-access-hzgcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:15:05 crc kubenswrapper[4894]: I1209 16:15:05.128871 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15be4c3e-4494-46ff-aca6-092e3eaef576-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "15be4c3e-4494-46ff-aca6-092e3eaef576" (UID: "15be4c3e-4494-46ff-aca6-092e3eaef576"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:15:05 crc kubenswrapper[4894]: I1209 16:15:05.216868 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hzgcx\" (UniqueName: \"kubernetes.io/projected/15be4c3e-4494-46ff-aca6-092e3eaef576-kube-api-access-hzgcx\") on node \"crc\" DevicePath \"\"" Dec 09 16:15:05 crc kubenswrapper[4894]: I1209 16:15:05.216909 4894 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/15be4c3e-4494-46ff-aca6-092e3eaef576-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 16:15:05 crc kubenswrapper[4894]: I1209 16:15:05.589385 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421615-fdjmw" event={"ID":"15be4c3e-4494-46ff-aca6-092e3eaef576","Type":"ContainerDied","Data":"21c0545d1886dcef4bc3a763e357e3b2614df5e4b6945fa2ecc95628609e5c8c"} Dec 09 16:15:05 crc kubenswrapper[4894]: I1209 16:15:05.589485 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="21c0545d1886dcef4bc3a763e357e3b2614df5e4b6945fa2ecc95628609e5c8c" Dec 09 16:15:05 crc kubenswrapper[4894]: I1209 16:15:05.589585 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421615-fdjmw" Dec 09 16:15:06 crc kubenswrapper[4894]: I1209 16:15:06.031089 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r"] Dec 09 16:15:06 crc kubenswrapper[4894]: I1209 16:15:06.041457 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421570-jcf7r"] Dec 09 16:15:06 crc kubenswrapper[4894]: I1209 16:15:06.140477 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f67460f9-bc45-47e0-b186-f66dea9a578c" path="/var/lib/kubelet/pods/f67460f9-bc45-47e0-b186-f66dea9a578c/volumes" Dec 09 16:15:57 crc kubenswrapper[4894]: I1209 16:15:57.371633 4894 scope.go:117] "RemoveContainer" containerID="d41ad5679ea1a20ef7453e9a31ff71dd7d5f8395803d542aa6e12f46dafa93f3" Dec 09 16:16:12 crc kubenswrapper[4894]: I1209 16:16:12.257457 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:16:12 crc kubenswrapper[4894]: I1209 16:16:12.258865 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:16:42 crc kubenswrapper[4894]: I1209 16:16:42.257901 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:16:42 crc kubenswrapper[4894]: I1209 16:16:42.258493 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:17:03 crc kubenswrapper[4894]: I1209 16:17:03.189523 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-4rsp8"] Dec 09 16:17:03 crc kubenswrapper[4894]: E1209 16:17:03.190799 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15be4c3e-4494-46ff-aca6-092e3eaef576" containerName="collect-profiles" Dec 09 16:17:03 crc kubenswrapper[4894]: I1209 16:17:03.190818 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="15be4c3e-4494-46ff-aca6-092e3eaef576" containerName="collect-profiles" Dec 09 16:17:03 crc kubenswrapper[4894]: I1209 16:17:03.191054 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="15be4c3e-4494-46ff-aca6-092e3eaef576" containerName="collect-profiles" Dec 09 16:17:03 crc kubenswrapper[4894]: I1209 16:17:03.196491 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4rsp8" Dec 09 16:17:03 crc kubenswrapper[4894]: I1209 16:17:03.213966 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4rsp8"] Dec 09 16:17:03 crc kubenswrapper[4894]: I1209 16:17:03.289232 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbe33793-56fc-4259-a706-535b373e61be-utilities\") pod \"redhat-operators-4rsp8\" (UID: \"dbe33793-56fc-4259-a706-535b373e61be\") " pod="openshift-marketplace/redhat-operators-4rsp8" Dec 09 16:17:03 crc kubenswrapper[4894]: I1209 16:17:03.289298 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mcfrk\" (UniqueName: \"kubernetes.io/projected/dbe33793-56fc-4259-a706-535b373e61be-kube-api-access-mcfrk\") pod \"redhat-operators-4rsp8\" (UID: \"dbe33793-56fc-4259-a706-535b373e61be\") " pod="openshift-marketplace/redhat-operators-4rsp8" Dec 09 16:17:03 crc kubenswrapper[4894]: I1209 16:17:03.289352 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbe33793-56fc-4259-a706-535b373e61be-catalog-content\") pod \"redhat-operators-4rsp8\" (UID: \"dbe33793-56fc-4259-a706-535b373e61be\") " pod="openshift-marketplace/redhat-operators-4rsp8" Dec 09 16:17:03 crc kubenswrapper[4894]: I1209 16:17:03.390652 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbe33793-56fc-4259-a706-535b373e61be-utilities\") pod \"redhat-operators-4rsp8\" (UID: \"dbe33793-56fc-4259-a706-535b373e61be\") " pod="openshift-marketplace/redhat-operators-4rsp8" Dec 09 16:17:03 crc kubenswrapper[4894]: I1209 16:17:03.390697 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mcfrk\" (UniqueName: \"kubernetes.io/projected/dbe33793-56fc-4259-a706-535b373e61be-kube-api-access-mcfrk\") pod \"redhat-operators-4rsp8\" (UID: \"dbe33793-56fc-4259-a706-535b373e61be\") " pod="openshift-marketplace/redhat-operators-4rsp8" Dec 09 16:17:03 crc kubenswrapper[4894]: I1209 16:17:03.391129 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbe33793-56fc-4259-a706-535b373e61be-catalog-content\") pod \"redhat-operators-4rsp8\" (UID: \"dbe33793-56fc-4259-a706-535b373e61be\") " pod="openshift-marketplace/redhat-operators-4rsp8" Dec 09 16:17:03 crc kubenswrapper[4894]: I1209 16:17:03.391252 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbe33793-56fc-4259-a706-535b373e61be-utilities\") pod \"redhat-operators-4rsp8\" (UID: \"dbe33793-56fc-4259-a706-535b373e61be\") " pod="openshift-marketplace/redhat-operators-4rsp8" Dec 09 16:17:03 crc kubenswrapper[4894]: I1209 16:17:03.391344 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbe33793-56fc-4259-a706-535b373e61be-catalog-content\") pod \"redhat-operators-4rsp8\" (UID: \"dbe33793-56fc-4259-a706-535b373e61be\") " pod="openshift-marketplace/redhat-operators-4rsp8" Dec 09 16:17:03 crc kubenswrapper[4894]: I1209 16:17:03.421680 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mcfrk\" (UniqueName: \"kubernetes.io/projected/dbe33793-56fc-4259-a706-535b373e61be-kube-api-access-mcfrk\") pod \"redhat-operators-4rsp8\" (UID: \"dbe33793-56fc-4259-a706-535b373e61be\") " pod="openshift-marketplace/redhat-operators-4rsp8" Dec 09 16:17:03 crc kubenswrapper[4894]: I1209 16:17:03.534871 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4rsp8" Dec 09 16:17:04 crc kubenswrapper[4894]: I1209 16:17:04.016956 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-4rsp8"] Dec 09 16:17:04 crc kubenswrapper[4894]: I1209 16:17:04.845266 4894 generic.go:334] "Generic (PLEG): container finished" podID="dbe33793-56fc-4259-a706-535b373e61be" containerID="c942fe4b260a66cca0d8cae1ef56dda8835107260ebdc03c4b63bc38b480f283" exitCode=0 Dec 09 16:17:04 crc kubenswrapper[4894]: I1209 16:17:04.845361 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4rsp8" event={"ID":"dbe33793-56fc-4259-a706-535b373e61be","Type":"ContainerDied","Data":"c942fe4b260a66cca0d8cae1ef56dda8835107260ebdc03c4b63bc38b480f283"} Dec 09 16:17:04 crc kubenswrapper[4894]: I1209 16:17:04.846418 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4rsp8" event={"ID":"dbe33793-56fc-4259-a706-535b373e61be","Type":"ContainerStarted","Data":"c35a3e568e100029169ac4ac1f49de8c8284c60c37d6eda3474317d082ee7315"} Dec 09 16:17:06 crc kubenswrapper[4894]: I1209 16:17:06.870425 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4rsp8" event={"ID":"dbe33793-56fc-4259-a706-535b373e61be","Type":"ContainerStarted","Data":"58eb277f3f4dbdb48d3f45ac6d947700d1161b16f17b9cfd45c234b54eec9ef0"} Dec 09 16:17:07 crc kubenswrapper[4894]: I1209 16:17:07.883205 4894 generic.go:334] "Generic (PLEG): container finished" podID="dbe33793-56fc-4259-a706-535b373e61be" containerID="58eb277f3f4dbdb48d3f45ac6d947700d1161b16f17b9cfd45c234b54eec9ef0" exitCode=0 Dec 09 16:17:07 crc kubenswrapper[4894]: I1209 16:17:07.883360 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4rsp8" event={"ID":"dbe33793-56fc-4259-a706-535b373e61be","Type":"ContainerDied","Data":"58eb277f3f4dbdb48d3f45ac6d947700d1161b16f17b9cfd45c234b54eec9ef0"} Dec 09 16:17:10 crc kubenswrapper[4894]: I1209 16:17:10.911707 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4rsp8" event={"ID":"dbe33793-56fc-4259-a706-535b373e61be","Type":"ContainerStarted","Data":"3008fe575e6caa711bc994f4d08c037cfa555b625f815a879b65fc04f5a1e516"} Dec 09 16:17:10 crc kubenswrapper[4894]: I1209 16:17:10.941529 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-4rsp8" podStartSLOduration=2.772315923 podStartE2EDuration="7.941511127s" podCreationTimestamp="2025-12-09 16:17:03 +0000 UTC" firstStartedPulling="2025-12-09 16:17:04.847534383 +0000 UTC m=+2719.166745052" lastFinishedPulling="2025-12-09 16:17:10.016729577 +0000 UTC m=+2724.335940256" observedRunningTime="2025-12-09 16:17:10.934902377 +0000 UTC m=+2725.254113046" watchObservedRunningTime="2025-12-09 16:17:10.941511127 +0000 UTC m=+2725.260721796" Dec 09 16:17:12 crc kubenswrapper[4894]: I1209 16:17:12.257378 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:17:12 crc kubenswrapper[4894]: I1209 16:17:12.257733 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:17:12 crc kubenswrapper[4894]: I1209 16:17:12.257789 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 16:17:12 crc kubenswrapper[4894]: I1209 16:17:12.258674 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"528a0efd27e59e2fb3d60ccf4dc2db0b0f33df50d99afe9391f928ded08456ba"} pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 16:17:12 crc kubenswrapper[4894]: I1209 16:17:12.258739 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" containerID="cri-o://528a0efd27e59e2fb3d60ccf4dc2db0b0f33df50d99afe9391f928ded08456ba" gracePeriod=600 Dec 09 16:17:12 crc kubenswrapper[4894]: I1209 16:17:12.933906 4894 generic.go:334] "Generic (PLEG): container finished" podID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerID="528a0efd27e59e2fb3d60ccf4dc2db0b0f33df50d99afe9391f928ded08456ba" exitCode=0 Dec 09 16:17:12 crc kubenswrapper[4894]: I1209 16:17:12.933947 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerDied","Data":"528a0efd27e59e2fb3d60ccf4dc2db0b0f33df50d99afe9391f928ded08456ba"} Dec 09 16:17:12 crc kubenswrapper[4894]: I1209 16:17:12.933978 4894 scope.go:117] "RemoveContainer" containerID="fa606814ffc61b3fa797b19f4610f3c5476ce61766f5167e8ebb805f79f94678" Dec 09 16:17:13 crc kubenswrapper[4894]: I1209 16:17:13.535739 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-4rsp8" Dec 09 16:17:13 crc kubenswrapper[4894]: I1209 16:17:13.536326 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-4rsp8" Dec 09 16:17:13 crc kubenswrapper[4894]: I1209 16:17:13.950317 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerStarted","Data":"8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86"} Dec 09 16:17:14 crc kubenswrapper[4894]: I1209 16:17:14.585564 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-4rsp8" podUID="dbe33793-56fc-4259-a706-535b373e61be" containerName="registry-server" probeResult="failure" output=< Dec 09 16:17:14 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 09 16:17:14 crc kubenswrapper[4894]: > Dec 09 16:17:23 crc kubenswrapper[4894]: I1209 16:17:23.599469 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-4rsp8" Dec 09 16:17:23 crc kubenswrapper[4894]: I1209 16:17:23.678665 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-4rsp8" Dec 09 16:17:23 crc kubenswrapper[4894]: I1209 16:17:23.839863 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4rsp8"] Dec 09 16:17:25 crc kubenswrapper[4894]: I1209 16:17:25.069315 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-4rsp8" podUID="dbe33793-56fc-4259-a706-535b373e61be" containerName="registry-server" containerID="cri-o://3008fe575e6caa711bc994f4d08c037cfa555b625f815a879b65fc04f5a1e516" gracePeriod=2 Dec 09 16:17:25 crc kubenswrapper[4894]: I1209 16:17:25.507562 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4rsp8" Dec 09 16:17:25 crc kubenswrapper[4894]: I1209 16:17:25.621952 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbe33793-56fc-4259-a706-535b373e61be-utilities\") pod \"dbe33793-56fc-4259-a706-535b373e61be\" (UID: \"dbe33793-56fc-4259-a706-535b373e61be\") " Dec 09 16:17:25 crc kubenswrapper[4894]: I1209 16:17:25.622006 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mcfrk\" (UniqueName: \"kubernetes.io/projected/dbe33793-56fc-4259-a706-535b373e61be-kube-api-access-mcfrk\") pod \"dbe33793-56fc-4259-a706-535b373e61be\" (UID: \"dbe33793-56fc-4259-a706-535b373e61be\") " Dec 09 16:17:25 crc kubenswrapper[4894]: I1209 16:17:25.622033 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbe33793-56fc-4259-a706-535b373e61be-catalog-content\") pod \"dbe33793-56fc-4259-a706-535b373e61be\" (UID: \"dbe33793-56fc-4259-a706-535b373e61be\") " Dec 09 16:17:25 crc kubenswrapper[4894]: I1209 16:17:25.622748 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbe33793-56fc-4259-a706-535b373e61be-utilities" (OuterVolumeSpecName: "utilities") pod "dbe33793-56fc-4259-a706-535b373e61be" (UID: "dbe33793-56fc-4259-a706-535b373e61be"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:17:25 crc kubenswrapper[4894]: I1209 16:17:25.628261 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dbe33793-56fc-4259-a706-535b373e61be-kube-api-access-mcfrk" (OuterVolumeSpecName: "kube-api-access-mcfrk") pod "dbe33793-56fc-4259-a706-535b373e61be" (UID: "dbe33793-56fc-4259-a706-535b373e61be"). InnerVolumeSpecName "kube-api-access-mcfrk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:17:25 crc kubenswrapper[4894]: I1209 16:17:25.724622 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dbe33793-56fc-4259-a706-535b373e61be-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:17:25 crc kubenswrapper[4894]: I1209 16:17:25.724686 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mcfrk\" (UniqueName: \"kubernetes.io/projected/dbe33793-56fc-4259-a706-535b373e61be-kube-api-access-mcfrk\") on node \"crc\" DevicePath \"\"" Dec 09 16:17:25 crc kubenswrapper[4894]: I1209 16:17:25.725037 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dbe33793-56fc-4259-a706-535b373e61be-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dbe33793-56fc-4259-a706-535b373e61be" (UID: "dbe33793-56fc-4259-a706-535b373e61be"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:17:25 crc kubenswrapper[4894]: I1209 16:17:25.825973 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dbe33793-56fc-4259-a706-535b373e61be-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:17:26 crc kubenswrapper[4894]: I1209 16:17:26.078304 4894 generic.go:334] "Generic (PLEG): container finished" podID="dbe33793-56fc-4259-a706-535b373e61be" containerID="3008fe575e6caa711bc994f4d08c037cfa555b625f815a879b65fc04f5a1e516" exitCode=0 Dec 09 16:17:26 crc kubenswrapper[4894]: I1209 16:17:26.078343 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4rsp8" event={"ID":"dbe33793-56fc-4259-a706-535b373e61be","Type":"ContainerDied","Data":"3008fe575e6caa711bc994f4d08c037cfa555b625f815a879b65fc04f5a1e516"} Dec 09 16:17:26 crc kubenswrapper[4894]: I1209 16:17:26.078369 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-4rsp8" event={"ID":"dbe33793-56fc-4259-a706-535b373e61be","Type":"ContainerDied","Data":"c35a3e568e100029169ac4ac1f49de8c8284c60c37d6eda3474317d082ee7315"} Dec 09 16:17:26 crc kubenswrapper[4894]: I1209 16:17:26.078368 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-4rsp8" Dec 09 16:17:26 crc kubenswrapper[4894]: I1209 16:17:26.078400 4894 scope.go:117] "RemoveContainer" containerID="3008fe575e6caa711bc994f4d08c037cfa555b625f815a879b65fc04f5a1e516" Dec 09 16:17:26 crc kubenswrapper[4894]: I1209 16:17:26.103456 4894 scope.go:117] "RemoveContainer" containerID="58eb277f3f4dbdb48d3f45ac6d947700d1161b16f17b9cfd45c234b54eec9ef0" Dec 09 16:17:26 crc kubenswrapper[4894]: I1209 16:17:26.117022 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-4rsp8"] Dec 09 16:17:26 crc kubenswrapper[4894]: I1209 16:17:26.118930 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-4rsp8"] Dec 09 16:17:26 crc kubenswrapper[4894]: I1209 16:17:26.147075 4894 scope.go:117] "RemoveContainer" containerID="c942fe4b260a66cca0d8cae1ef56dda8835107260ebdc03c4b63bc38b480f283" Dec 09 16:17:26 crc kubenswrapper[4894]: I1209 16:17:26.166240 4894 scope.go:117] "RemoveContainer" containerID="3008fe575e6caa711bc994f4d08c037cfa555b625f815a879b65fc04f5a1e516" Dec 09 16:17:26 crc kubenswrapper[4894]: E1209 16:17:26.166584 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3008fe575e6caa711bc994f4d08c037cfa555b625f815a879b65fc04f5a1e516\": container with ID starting with 3008fe575e6caa711bc994f4d08c037cfa555b625f815a879b65fc04f5a1e516 not found: ID does not exist" containerID="3008fe575e6caa711bc994f4d08c037cfa555b625f815a879b65fc04f5a1e516" Dec 09 16:17:26 crc kubenswrapper[4894]: I1209 16:17:26.166620 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3008fe575e6caa711bc994f4d08c037cfa555b625f815a879b65fc04f5a1e516"} err="failed to get container status \"3008fe575e6caa711bc994f4d08c037cfa555b625f815a879b65fc04f5a1e516\": rpc error: code = NotFound desc = could not find container \"3008fe575e6caa711bc994f4d08c037cfa555b625f815a879b65fc04f5a1e516\": container with ID starting with 3008fe575e6caa711bc994f4d08c037cfa555b625f815a879b65fc04f5a1e516 not found: ID does not exist" Dec 09 16:17:26 crc kubenswrapper[4894]: I1209 16:17:26.166675 4894 scope.go:117] "RemoveContainer" containerID="58eb277f3f4dbdb48d3f45ac6d947700d1161b16f17b9cfd45c234b54eec9ef0" Dec 09 16:17:26 crc kubenswrapper[4894]: E1209 16:17:26.166882 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"58eb277f3f4dbdb48d3f45ac6d947700d1161b16f17b9cfd45c234b54eec9ef0\": container with ID starting with 58eb277f3f4dbdb48d3f45ac6d947700d1161b16f17b9cfd45c234b54eec9ef0 not found: ID does not exist" containerID="58eb277f3f4dbdb48d3f45ac6d947700d1161b16f17b9cfd45c234b54eec9ef0" Dec 09 16:17:26 crc kubenswrapper[4894]: I1209 16:17:26.166906 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"58eb277f3f4dbdb48d3f45ac6d947700d1161b16f17b9cfd45c234b54eec9ef0"} err="failed to get container status \"58eb277f3f4dbdb48d3f45ac6d947700d1161b16f17b9cfd45c234b54eec9ef0\": rpc error: code = NotFound desc = could not find container \"58eb277f3f4dbdb48d3f45ac6d947700d1161b16f17b9cfd45c234b54eec9ef0\": container with ID starting with 58eb277f3f4dbdb48d3f45ac6d947700d1161b16f17b9cfd45c234b54eec9ef0 not found: ID does not exist" Dec 09 16:17:26 crc kubenswrapper[4894]: I1209 16:17:26.166922 4894 scope.go:117] "RemoveContainer" containerID="c942fe4b260a66cca0d8cae1ef56dda8835107260ebdc03c4b63bc38b480f283" Dec 09 16:17:26 crc kubenswrapper[4894]: E1209 16:17:26.167069 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c942fe4b260a66cca0d8cae1ef56dda8835107260ebdc03c4b63bc38b480f283\": container with ID starting with c942fe4b260a66cca0d8cae1ef56dda8835107260ebdc03c4b63bc38b480f283 not found: ID does not exist" containerID="c942fe4b260a66cca0d8cae1ef56dda8835107260ebdc03c4b63bc38b480f283" Dec 09 16:17:26 crc kubenswrapper[4894]: I1209 16:17:26.167090 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c942fe4b260a66cca0d8cae1ef56dda8835107260ebdc03c4b63bc38b480f283"} err="failed to get container status \"c942fe4b260a66cca0d8cae1ef56dda8835107260ebdc03c4b63bc38b480f283\": rpc error: code = NotFound desc = could not find container \"c942fe4b260a66cca0d8cae1ef56dda8835107260ebdc03c4b63bc38b480f283\": container with ID starting with c942fe4b260a66cca0d8cae1ef56dda8835107260ebdc03c4b63bc38b480f283 not found: ID does not exist" Dec 09 16:17:28 crc kubenswrapper[4894]: I1209 16:17:28.120767 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dbe33793-56fc-4259-a706-535b373e61be" path="/var/lib/kubelet/pods/dbe33793-56fc-4259-a706-535b373e61be/volumes" Dec 09 16:18:43 crc kubenswrapper[4894]: I1209 16:18:43.920936 4894 generic.go:334] "Generic (PLEG): container finished" podID="f50d2ea8-df93-445e-9ce0-f5c9d4077cc1" containerID="7cac2f4617590ede70e797c27bfc2383a299d4c01c1a66edba1153d65296f729" exitCode=0 Dec 09 16:18:43 crc kubenswrapper[4894]: I1209 16:18:43.921008 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" event={"ID":"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1","Type":"ContainerDied","Data":"7cac2f4617590ede70e797c27bfc2383a299d4c01c1a66edba1153d65296f729"} Dec 09 16:18:45 crc kubenswrapper[4894]: I1209 16:18:45.367649 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" Dec 09 16:18:45 crc kubenswrapper[4894]: I1209 16:18:45.499906 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-libvirt-secret-0\") pod \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " Dec 09 16:18:45 crc kubenswrapper[4894]: I1209 16:18:45.500323 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-ssh-key\") pod \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " Dec 09 16:18:45 crc kubenswrapper[4894]: I1209 16:18:45.500363 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-libvirt-combined-ca-bundle\") pod \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " Dec 09 16:18:45 crc kubenswrapper[4894]: I1209 16:18:45.500409 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-inventory\") pod \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " Dec 09 16:18:45 crc kubenswrapper[4894]: I1209 16:18:45.500435 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjlv8\" (UniqueName: \"kubernetes.io/projected/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-kube-api-access-sjlv8\") pod \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " Dec 09 16:18:45 crc kubenswrapper[4894]: I1209 16:18:45.500451 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-ceph\") pod \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\" (UID: \"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1\") " Dec 09 16:18:45 crc kubenswrapper[4894]: I1209 16:18:45.506425 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-libvirt-combined-ca-bundle" (OuterVolumeSpecName: "libvirt-combined-ca-bundle") pod "f50d2ea8-df93-445e-9ce0-f5c9d4077cc1" (UID: "f50d2ea8-df93-445e-9ce0-f5c9d4077cc1"). InnerVolumeSpecName "libvirt-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:18:45 crc kubenswrapper[4894]: I1209 16:18:45.507951 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-kube-api-access-sjlv8" (OuterVolumeSpecName: "kube-api-access-sjlv8") pod "f50d2ea8-df93-445e-9ce0-f5c9d4077cc1" (UID: "f50d2ea8-df93-445e-9ce0-f5c9d4077cc1"). InnerVolumeSpecName "kube-api-access-sjlv8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:18:45 crc kubenswrapper[4894]: I1209 16:18:45.514942 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-ceph" (OuterVolumeSpecName: "ceph") pod "f50d2ea8-df93-445e-9ce0-f5c9d4077cc1" (UID: "f50d2ea8-df93-445e-9ce0-f5c9d4077cc1"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:18:45 crc kubenswrapper[4894]: I1209 16:18:45.539553 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "f50d2ea8-df93-445e-9ce0-f5c9d4077cc1" (UID: "f50d2ea8-df93-445e-9ce0-f5c9d4077cc1"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:18:45 crc kubenswrapper[4894]: I1209 16:18:45.552960 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-libvirt-secret-0" (OuterVolumeSpecName: "libvirt-secret-0") pod "f50d2ea8-df93-445e-9ce0-f5c9d4077cc1" (UID: "f50d2ea8-df93-445e-9ce0-f5c9d4077cc1"). InnerVolumeSpecName "libvirt-secret-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:18:45 crc kubenswrapper[4894]: I1209 16:18:45.557507 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-inventory" (OuterVolumeSpecName: "inventory") pod "f50d2ea8-df93-445e-9ce0-f5c9d4077cc1" (UID: "f50d2ea8-df93-445e-9ce0-f5c9d4077cc1"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:18:45 crc kubenswrapper[4894]: I1209 16:18:45.603067 4894 reconciler_common.go:293] "Volume detached for volume \"libvirt-secret-0\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-libvirt-secret-0\") on node \"crc\" DevicePath \"\"" Dec 09 16:18:45 crc kubenswrapper[4894]: I1209 16:18:45.603098 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:18:45 crc kubenswrapper[4894]: I1209 16:18:45.603110 4894 reconciler_common.go:293] "Volume detached for volume \"libvirt-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-libvirt-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:18:45 crc kubenswrapper[4894]: I1209 16:18:45.603125 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 16:18:45 crc kubenswrapper[4894]: I1209 16:18:45.603136 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjlv8\" (UniqueName: \"kubernetes.io/projected/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-kube-api-access-sjlv8\") on node \"crc\" DevicePath \"\"" Dec 09 16:18:45 crc kubenswrapper[4894]: I1209 16:18:45.603147 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f50d2ea8-df93-445e-9ce0-f5c9d4077cc1-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:18:45 crc kubenswrapper[4894]: I1209 16:18:45.949463 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" event={"ID":"f50d2ea8-df93-445e-9ce0-f5c9d4077cc1","Type":"ContainerDied","Data":"c0ba307bd08f307f88d2d21f9ea10c64a25718a2a0f1cd76d5882d7936b1d24e"} Dec 09 16:18:45 crc kubenswrapper[4894]: I1209 16:18:45.949547 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c0ba307bd08f307f88d2d21f9ea10c64a25718a2a0f1cd76d5882d7936b1d24e" Dec 09 16:18:45 crc kubenswrapper[4894]: I1209 16:18:45.949679 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.087936 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29"] Dec 09 16:18:46 crc kubenswrapper[4894]: E1209 16:18:46.089021 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbe33793-56fc-4259-a706-535b373e61be" containerName="registry-server" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.089230 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbe33793-56fc-4259-a706-535b373e61be" containerName="registry-server" Dec 09 16:18:46 crc kubenswrapper[4894]: E1209 16:18:46.089437 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbe33793-56fc-4259-a706-535b373e61be" containerName="extract-utilities" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.089568 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbe33793-56fc-4259-a706-535b373e61be" containerName="extract-utilities" Dec 09 16:18:46 crc kubenswrapper[4894]: E1209 16:18:46.089806 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dbe33793-56fc-4259-a706-535b373e61be" containerName="extract-content" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.089952 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="dbe33793-56fc-4259-a706-535b373e61be" containerName="extract-content" Dec 09 16:18:46 crc kubenswrapper[4894]: E1209 16:18:46.090090 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f50d2ea8-df93-445e-9ce0-f5c9d4077cc1" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.090223 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f50d2ea8-df93-445e-9ce0-f5c9d4077cc1" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.090740 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f50d2ea8-df93-445e-9ce0-f5c9d4077cc1" containerName="libvirt-edpm-deployment-openstack-edpm-ipam" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.090954 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="dbe33793-56fc-4259-a706-535b373e61be" containerName="registry-server" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.092213 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.094680 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-aee-default-env" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.095401 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-compute-config" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.095699 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.095722 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.095752 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"nova-extra-config" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.095769 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ceph-nova" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.097749 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplanenodeset-openstack-edpm-ipam" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.098017 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-migration-ssh-key" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.098236 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-edpm-ipam-dockercfg-75rjx" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.099270 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29"] Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.113983 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.114030 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.114065 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.114110 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9625f994-137d-4957-a9f9-40bf22bf832b-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.114145 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pm8nx\" (UniqueName: \"kubernetes.io/projected/9625f994-137d-4957-a9f9-40bf22bf832b-kube-api-access-pm8nx\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.114173 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.114213 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.114266 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.114284 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.114330 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.114349 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/9625f994-137d-4957-a9f9-40bf22bf832b-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.216305 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9625f994-137d-4957-a9f9-40bf22bf832b-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.216412 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pm8nx\" (UniqueName: \"kubernetes.io/projected/9625f994-137d-4957-a9f9-40bf22bf832b-kube-api-access-pm8nx\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.216478 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.216556 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.216718 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.216761 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.216836 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.216868 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/9625f994-137d-4957-a9f9-40bf22bf832b-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.216909 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.216958 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.217161 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.217947 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9625f994-137d-4957-a9f9-40bf22bf832b-nova-extra-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.222609 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-cell1-compute-config-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.223602 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-migration-ssh-key-1\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.223660 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-migration-ssh-key-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.226217 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-custom-ceph-combined-ca-bundle\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.226288 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-ssh-key\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.226738 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-inventory\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.227327 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-ceph\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.228316 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-cell1-compute-config-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.228533 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/9625f994-137d-4957-a9f9-40bf22bf832b-ceph-nova-0\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.240738 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pm8nx\" (UniqueName: \"kubernetes.io/projected/9625f994-137d-4957-a9f9-40bf22bf832b-kube-api-access-pm8nx\") pod \"nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:46 crc kubenswrapper[4894]: I1209 16:18:46.419698 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:18:47 crc kubenswrapper[4894]: I1209 16:18:47.092893 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29"] Dec 09 16:18:47 crc kubenswrapper[4894]: I1209 16:18:47.971095 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" event={"ID":"9625f994-137d-4957-a9f9-40bf22bf832b","Type":"ContainerStarted","Data":"523f77a99e598917506b6d282f93c032ea65adbf4e2f71edd6836d8ab10b5bf0"} Dec 09 16:18:48 crc kubenswrapper[4894]: I1209 16:18:48.985019 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" event={"ID":"9625f994-137d-4957-a9f9-40bf22bf832b","Type":"ContainerStarted","Data":"bf1ef71345d27d6192671dad28d741115ed4f1b805521d72820615af465a693b"} Dec 09 16:18:49 crc kubenswrapper[4894]: I1209 16:18:49.015465 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" podStartSLOduration=2.338044682 podStartE2EDuration="3.015449282s" podCreationTimestamp="2025-12-09 16:18:46 +0000 UTC" firstStartedPulling="2025-12-09 16:18:47.096276235 +0000 UTC m=+2821.415486934" lastFinishedPulling="2025-12-09 16:18:47.773680855 +0000 UTC m=+2822.092891534" observedRunningTime="2025-12-09 16:18:49.010782375 +0000 UTC m=+2823.329993054" watchObservedRunningTime="2025-12-09 16:18:49.015449282 +0000 UTC m=+2823.334659951" Dec 09 16:19:42 crc kubenswrapper[4894]: I1209 16:19:42.257793 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:19:42 crc kubenswrapper[4894]: I1209 16:19:42.258541 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:20:12 crc kubenswrapper[4894]: I1209 16:20:12.257801 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:20:12 crc kubenswrapper[4894]: I1209 16:20:12.258902 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:20:16 crc kubenswrapper[4894]: I1209 16:20:16.585475 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hzrjp"] Dec 09 16:20:16 crc kubenswrapper[4894]: I1209 16:20:16.588409 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hzrjp" Dec 09 16:20:16 crc kubenswrapper[4894]: I1209 16:20:16.605905 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hzrjp"] Dec 09 16:20:16 crc kubenswrapper[4894]: I1209 16:20:16.663186 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c8xjd\" (UniqueName: \"kubernetes.io/projected/3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0-kube-api-access-c8xjd\") pod \"community-operators-hzrjp\" (UID: \"3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0\") " pod="openshift-marketplace/community-operators-hzrjp" Dec 09 16:20:16 crc kubenswrapper[4894]: I1209 16:20:16.663270 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0-utilities\") pod \"community-operators-hzrjp\" (UID: \"3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0\") " pod="openshift-marketplace/community-operators-hzrjp" Dec 09 16:20:16 crc kubenswrapper[4894]: I1209 16:20:16.663385 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0-catalog-content\") pod \"community-operators-hzrjp\" (UID: \"3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0\") " pod="openshift-marketplace/community-operators-hzrjp" Dec 09 16:20:16 crc kubenswrapper[4894]: I1209 16:20:16.765307 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c8xjd\" (UniqueName: \"kubernetes.io/projected/3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0-kube-api-access-c8xjd\") pod \"community-operators-hzrjp\" (UID: \"3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0\") " pod="openshift-marketplace/community-operators-hzrjp" Dec 09 16:20:16 crc kubenswrapper[4894]: I1209 16:20:16.765377 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0-utilities\") pod \"community-operators-hzrjp\" (UID: \"3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0\") " pod="openshift-marketplace/community-operators-hzrjp" Dec 09 16:20:16 crc kubenswrapper[4894]: I1209 16:20:16.765538 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0-catalog-content\") pod \"community-operators-hzrjp\" (UID: \"3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0\") " pod="openshift-marketplace/community-operators-hzrjp" Dec 09 16:20:16 crc kubenswrapper[4894]: I1209 16:20:16.767831 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0-utilities\") pod \"community-operators-hzrjp\" (UID: \"3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0\") " pod="openshift-marketplace/community-operators-hzrjp" Dec 09 16:20:16 crc kubenswrapper[4894]: I1209 16:20:16.768212 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0-catalog-content\") pod \"community-operators-hzrjp\" (UID: \"3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0\") " pod="openshift-marketplace/community-operators-hzrjp" Dec 09 16:20:16 crc kubenswrapper[4894]: I1209 16:20:16.801375 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c8xjd\" (UniqueName: \"kubernetes.io/projected/3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0-kube-api-access-c8xjd\") pod \"community-operators-hzrjp\" (UID: \"3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0\") " pod="openshift-marketplace/community-operators-hzrjp" Dec 09 16:20:16 crc kubenswrapper[4894]: I1209 16:20:16.946242 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hzrjp" Dec 09 16:20:17 crc kubenswrapper[4894]: I1209 16:20:17.438703 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hzrjp"] Dec 09 16:20:17 crc kubenswrapper[4894]: I1209 16:20:17.922191 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzrjp" event={"ID":"3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0","Type":"ContainerStarted","Data":"66a437cd1b381757b3b287070065ce13d0d8c33b31a3c8a470b06f03d5677351"} Dec 09 16:20:18 crc kubenswrapper[4894]: I1209 16:20:18.930588 4894 generic.go:334] "Generic (PLEG): container finished" podID="3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0" containerID="bbcd88ac5398d7f72edcaca84dbea3e4fd5044be3c9d15878b2a45264f95792f" exitCode=0 Dec 09 16:20:18 crc kubenswrapper[4894]: I1209 16:20:18.930684 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzrjp" event={"ID":"3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0","Type":"ContainerDied","Data":"bbcd88ac5398d7f72edcaca84dbea3e4fd5044be3c9d15878b2a45264f95792f"} Dec 09 16:20:18 crc kubenswrapper[4894]: I1209 16:20:18.933232 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 16:20:20 crc kubenswrapper[4894]: I1209 16:20:20.956532 4894 generic.go:334] "Generic (PLEG): container finished" podID="3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0" containerID="003e5857c5aba3c7d761d7565600819ce06a9e8238444b369dde08b1c579f6c3" exitCode=0 Dec 09 16:20:20 crc kubenswrapper[4894]: I1209 16:20:20.956698 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzrjp" event={"ID":"3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0","Type":"ContainerDied","Data":"003e5857c5aba3c7d761d7565600819ce06a9e8238444b369dde08b1c579f6c3"} Dec 09 16:20:21 crc kubenswrapper[4894]: I1209 16:20:21.971253 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzrjp" event={"ID":"3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0","Type":"ContainerStarted","Data":"d4820d875df61ee9ea0517d1b7194bf8482cff17b60507debee324bcb46ff7dd"} Dec 09 16:20:21 crc kubenswrapper[4894]: I1209 16:20:21.997363 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hzrjp" podStartSLOduration=3.50179031 podStartE2EDuration="5.997346609s" podCreationTimestamp="2025-12-09 16:20:16 +0000 UTC" firstStartedPulling="2025-12-09 16:20:18.932704674 +0000 UTC m=+2913.251915393" lastFinishedPulling="2025-12-09 16:20:21.428260983 +0000 UTC m=+2915.747471692" observedRunningTime="2025-12-09 16:20:21.990330849 +0000 UTC m=+2916.309541528" watchObservedRunningTime="2025-12-09 16:20:21.997346609 +0000 UTC m=+2916.316557278" Dec 09 16:20:26 crc kubenswrapper[4894]: I1209 16:20:26.947033 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hzrjp" Dec 09 16:20:26 crc kubenswrapper[4894]: I1209 16:20:26.947542 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hzrjp" Dec 09 16:20:27 crc kubenswrapper[4894]: I1209 16:20:27.027879 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hzrjp" Dec 09 16:20:27 crc kubenswrapper[4894]: I1209 16:20:27.109716 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hzrjp" Dec 09 16:20:27 crc kubenswrapper[4894]: I1209 16:20:27.287025 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hzrjp"] Dec 09 16:20:29 crc kubenswrapper[4894]: I1209 16:20:29.054150 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hzrjp" podUID="3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0" containerName="registry-server" containerID="cri-o://d4820d875df61ee9ea0517d1b7194bf8482cff17b60507debee324bcb46ff7dd" gracePeriod=2 Dec 09 16:20:29 crc kubenswrapper[4894]: I1209 16:20:29.637952 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hzrjp" Dec 09 16:20:29 crc kubenswrapper[4894]: I1209 16:20:29.748123 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0-catalog-content\") pod \"3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0\" (UID: \"3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0\") " Dec 09 16:20:29 crc kubenswrapper[4894]: I1209 16:20:29.748171 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c8xjd\" (UniqueName: \"kubernetes.io/projected/3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0-kube-api-access-c8xjd\") pod \"3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0\" (UID: \"3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0\") " Dec 09 16:20:29 crc kubenswrapper[4894]: I1209 16:20:29.748395 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0-utilities\") pod \"3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0\" (UID: \"3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0\") " Dec 09 16:20:29 crc kubenswrapper[4894]: I1209 16:20:29.749075 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0-utilities" (OuterVolumeSpecName: "utilities") pod "3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0" (UID: "3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:20:29 crc kubenswrapper[4894]: I1209 16:20:29.767785 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0-kube-api-access-c8xjd" (OuterVolumeSpecName: "kube-api-access-c8xjd") pod "3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0" (UID: "3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0"). InnerVolumeSpecName "kube-api-access-c8xjd". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:20:29 crc kubenswrapper[4894]: I1209 16:20:29.823359 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0" (UID: "3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:20:29 crc kubenswrapper[4894]: I1209 16:20:29.851124 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:20:29 crc kubenswrapper[4894]: I1209 16:20:29.851171 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c8xjd\" (UniqueName: \"kubernetes.io/projected/3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0-kube-api-access-c8xjd\") on node \"crc\" DevicePath \"\"" Dec 09 16:20:29 crc kubenswrapper[4894]: I1209 16:20:29.851192 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:20:30 crc kubenswrapper[4894]: I1209 16:20:30.071413 4894 generic.go:334] "Generic (PLEG): container finished" podID="3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0" containerID="d4820d875df61ee9ea0517d1b7194bf8482cff17b60507debee324bcb46ff7dd" exitCode=0 Dec 09 16:20:30 crc kubenswrapper[4894]: I1209 16:20:30.071465 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzrjp" event={"ID":"3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0","Type":"ContainerDied","Data":"d4820d875df61ee9ea0517d1b7194bf8482cff17b60507debee324bcb46ff7dd"} Dec 09 16:20:30 crc kubenswrapper[4894]: I1209 16:20:30.071495 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hzrjp" event={"ID":"3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0","Type":"ContainerDied","Data":"66a437cd1b381757b3b287070065ce13d0d8c33b31a3c8a470b06f03d5677351"} Dec 09 16:20:30 crc kubenswrapper[4894]: I1209 16:20:30.071518 4894 scope.go:117] "RemoveContainer" containerID="d4820d875df61ee9ea0517d1b7194bf8482cff17b60507debee324bcb46ff7dd" Dec 09 16:20:30 crc kubenswrapper[4894]: I1209 16:20:30.071686 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hzrjp" Dec 09 16:20:30 crc kubenswrapper[4894]: I1209 16:20:30.114139 4894 scope.go:117] "RemoveContainer" containerID="003e5857c5aba3c7d761d7565600819ce06a9e8238444b369dde08b1c579f6c3" Dec 09 16:20:30 crc kubenswrapper[4894]: I1209 16:20:30.127845 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hzrjp"] Dec 09 16:20:30 crc kubenswrapper[4894]: I1209 16:20:30.134897 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hzrjp"] Dec 09 16:20:30 crc kubenswrapper[4894]: I1209 16:20:30.155596 4894 scope.go:117] "RemoveContainer" containerID="bbcd88ac5398d7f72edcaca84dbea3e4fd5044be3c9d15878b2a45264f95792f" Dec 09 16:20:30 crc kubenswrapper[4894]: I1209 16:20:30.191913 4894 scope.go:117] "RemoveContainer" containerID="d4820d875df61ee9ea0517d1b7194bf8482cff17b60507debee324bcb46ff7dd" Dec 09 16:20:30 crc kubenswrapper[4894]: E1209 16:20:30.192489 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4820d875df61ee9ea0517d1b7194bf8482cff17b60507debee324bcb46ff7dd\": container with ID starting with d4820d875df61ee9ea0517d1b7194bf8482cff17b60507debee324bcb46ff7dd not found: ID does not exist" containerID="d4820d875df61ee9ea0517d1b7194bf8482cff17b60507debee324bcb46ff7dd" Dec 09 16:20:30 crc kubenswrapper[4894]: I1209 16:20:30.192557 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4820d875df61ee9ea0517d1b7194bf8482cff17b60507debee324bcb46ff7dd"} err="failed to get container status \"d4820d875df61ee9ea0517d1b7194bf8482cff17b60507debee324bcb46ff7dd\": rpc error: code = NotFound desc = could not find container \"d4820d875df61ee9ea0517d1b7194bf8482cff17b60507debee324bcb46ff7dd\": container with ID starting with d4820d875df61ee9ea0517d1b7194bf8482cff17b60507debee324bcb46ff7dd not found: ID does not exist" Dec 09 16:20:30 crc kubenswrapper[4894]: I1209 16:20:30.192589 4894 scope.go:117] "RemoveContainer" containerID="003e5857c5aba3c7d761d7565600819ce06a9e8238444b369dde08b1c579f6c3" Dec 09 16:20:30 crc kubenswrapper[4894]: E1209 16:20:30.193038 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"003e5857c5aba3c7d761d7565600819ce06a9e8238444b369dde08b1c579f6c3\": container with ID starting with 003e5857c5aba3c7d761d7565600819ce06a9e8238444b369dde08b1c579f6c3 not found: ID does not exist" containerID="003e5857c5aba3c7d761d7565600819ce06a9e8238444b369dde08b1c579f6c3" Dec 09 16:20:30 crc kubenswrapper[4894]: I1209 16:20:30.193164 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"003e5857c5aba3c7d761d7565600819ce06a9e8238444b369dde08b1c579f6c3"} err="failed to get container status \"003e5857c5aba3c7d761d7565600819ce06a9e8238444b369dde08b1c579f6c3\": rpc error: code = NotFound desc = could not find container \"003e5857c5aba3c7d761d7565600819ce06a9e8238444b369dde08b1c579f6c3\": container with ID starting with 003e5857c5aba3c7d761d7565600819ce06a9e8238444b369dde08b1c579f6c3 not found: ID does not exist" Dec 09 16:20:30 crc kubenswrapper[4894]: I1209 16:20:30.193195 4894 scope.go:117] "RemoveContainer" containerID="bbcd88ac5398d7f72edcaca84dbea3e4fd5044be3c9d15878b2a45264f95792f" Dec 09 16:20:30 crc kubenswrapper[4894]: E1209 16:20:30.193731 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbcd88ac5398d7f72edcaca84dbea3e4fd5044be3c9d15878b2a45264f95792f\": container with ID starting with bbcd88ac5398d7f72edcaca84dbea3e4fd5044be3c9d15878b2a45264f95792f not found: ID does not exist" containerID="bbcd88ac5398d7f72edcaca84dbea3e4fd5044be3c9d15878b2a45264f95792f" Dec 09 16:20:30 crc kubenswrapper[4894]: I1209 16:20:30.193782 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbcd88ac5398d7f72edcaca84dbea3e4fd5044be3c9d15878b2a45264f95792f"} err="failed to get container status \"bbcd88ac5398d7f72edcaca84dbea3e4fd5044be3c9d15878b2a45264f95792f\": rpc error: code = NotFound desc = could not find container \"bbcd88ac5398d7f72edcaca84dbea3e4fd5044be3c9d15878b2a45264f95792f\": container with ID starting with bbcd88ac5398d7f72edcaca84dbea3e4fd5044be3c9d15878b2a45264f95792f not found: ID does not exist" Dec 09 16:20:32 crc kubenswrapper[4894]: I1209 16:20:32.129795 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0" path="/var/lib/kubelet/pods/3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0/volumes" Dec 09 16:20:42 crc kubenswrapper[4894]: I1209 16:20:42.257779 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:20:42 crc kubenswrapper[4894]: I1209 16:20:42.258374 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:20:42 crc kubenswrapper[4894]: I1209 16:20:42.258414 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 16:20:42 crc kubenswrapper[4894]: I1209 16:20:42.259235 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86"} pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 16:20:42 crc kubenswrapper[4894]: I1209 16:20:42.259302 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" containerID="cri-o://8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" gracePeriod=600 Dec 09 16:20:42 crc kubenswrapper[4894]: E1209 16:20:42.393421 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:20:43 crc kubenswrapper[4894]: I1209 16:20:43.231447 4894 generic.go:334] "Generic (PLEG): container finished" podID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" exitCode=0 Dec 09 16:20:43 crc kubenswrapper[4894]: I1209 16:20:43.231505 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerDied","Data":"8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86"} Dec 09 16:20:43 crc kubenswrapper[4894]: I1209 16:20:43.231557 4894 scope.go:117] "RemoveContainer" containerID="528a0efd27e59e2fb3d60ccf4dc2db0b0f33df50d99afe9391f928ded08456ba" Dec 09 16:20:43 crc kubenswrapper[4894]: I1209 16:20:43.232180 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:20:43 crc kubenswrapper[4894]: E1209 16:20:43.232538 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:20:58 crc kubenswrapper[4894]: I1209 16:20:58.107057 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:20:58 crc kubenswrapper[4894]: E1209 16:20:58.107892 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:21:10 crc kubenswrapper[4894]: I1209 16:21:10.106739 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:21:10 crc kubenswrapper[4894]: E1209 16:21:10.108022 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:21:22 crc kubenswrapper[4894]: I1209 16:21:22.106983 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:21:22 crc kubenswrapper[4894]: E1209 16:21:22.107784 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:21:33 crc kubenswrapper[4894]: I1209 16:21:33.106352 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:21:33 crc kubenswrapper[4894]: E1209 16:21:33.107371 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:21:36 crc kubenswrapper[4894]: I1209 16:21:36.703957 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lnckw"] Dec 09 16:21:36 crc kubenswrapper[4894]: E1209 16:21:36.707113 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0" containerName="extract-utilities" Dec 09 16:21:36 crc kubenswrapper[4894]: I1209 16:21:36.707171 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0" containerName="extract-utilities" Dec 09 16:21:36 crc kubenswrapper[4894]: E1209 16:21:36.707237 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0" containerName="extract-content" Dec 09 16:21:36 crc kubenswrapper[4894]: I1209 16:21:36.707252 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0" containerName="extract-content" Dec 09 16:21:36 crc kubenswrapper[4894]: E1209 16:21:36.707287 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0" containerName="registry-server" Dec 09 16:21:36 crc kubenswrapper[4894]: I1209 16:21:36.707299 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0" containerName="registry-server" Dec 09 16:21:36 crc kubenswrapper[4894]: I1209 16:21:36.707827 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="3a86e2fa-e621-434d-8bc1-9d2bcf0b2ec0" containerName="registry-server" Dec 09 16:21:36 crc kubenswrapper[4894]: I1209 16:21:36.715894 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnckw" Dec 09 16:21:36 crc kubenswrapper[4894]: I1209 16:21:36.725170 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnckw"] Dec 09 16:21:36 crc kubenswrapper[4894]: I1209 16:21:36.814174 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da7c493c-ccdf-4066-b3d8-246326d48806-utilities\") pod \"redhat-marketplace-lnckw\" (UID: \"da7c493c-ccdf-4066-b3d8-246326d48806\") " pod="openshift-marketplace/redhat-marketplace-lnckw" Dec 09 16:21:36 crc kubenswrapper[4894]: I1209 16:21:36.814450 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nz77s\" (UniqueName: \"kubernetes.io/projected/da7c493c-ccdf-4066-b3d8-246326d48806-kube-api-access-nz77s\") pod \"redhat-marketplace-lnckw\" (UID: \"da7c493c-ccdf-4066-b3d8-246326d48806\") " pod="openshift-marketplace/redhat-marketplace-lnckw" Dec 09 16:21:36 crc kubenswrapper[4894]: I1209 16:21:36.814611 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da7c493c-ccdf-4066-b3d8-246326d48806-catalog-content\") pod \"redhat-marketplace-lnckw\" (UID: \"da7c493c-ccdf-4066-b3d8-246326d48806\") " pod="openshift-marketplace/redhat-marketplace-lnckw" Dec 09 16:21:36 crc kubenswrapper[4894]: I1209 16:21:36.916216 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da7c493c-ccdf-4066-b3d8-246326d48806-utilities\") pod \"redhat-marketplace-lnckw\" (UID: \"da7c493c-ccdf-4066-b3d8-246326d48806\") " pod="openshift-marketplace/redhat-marketplace-lnckw" Dec 09 16:21:36 crc kubenswrapper[4894]: I1209 16:21:36.916258 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nz77s\" (UniqueName: \"kubernetes.io/projected/da7c493c-ccdf-4066-b3d8-246326d48806-kube-api-access-nz77s\") pod \"redhat-marketplace-lnckw\" (UID: \"da7c493c-ccdf-4066-b3d8-246326d48806\") " pod="openshift-marketplace/redhat-marketplace-lnckw" Dec 09 16:21:36 crc kubenswrapper[4894]: I1209 16:21:36.916304 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da7c493c-ccdf-4066-b3d8-246326d48806-catalog-content\") pod \"redhat-marketplace-lnckw\" (UID: \"da7c493c-ccdf-4066-b3d8-246326d48806\") " pod="openshift-marketplace/redhat-marketplace-lnckw" Dec 09 16:21:36 crc kubenswrapper[4894]: I1209 16:21:36.916698 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da7c493c-ccdf-4066-b3d8-246326d48806-utilities\") pod \"redhat-marketplace-lnckw\" (UID: \"da7c493c-ccdf-4066-b3d8-246326d48806\") " pod="openshift-marketplace/redhat-marketplace-lnckw" Dec 09 16:21:36 crc kubenswrapper[4894]: I1209 16:21:36.916761 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da7c493c-ccdf-4066-b3d8-246326d48806-catalog-content\") pod \"redhat-marketplace-lnckw\" (UID: \"da7c493c-ccdf-4066-b3d8-246326d48806\") " pod="openshift-marketplace/redhat-marketplace-lnckw" Dec 09 16:21:36 crc kubenswrapper[4894]: I1209 16:21:36.939358 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nz77s\" (UniqueName: \"kubernetes.io/projected/da7c493c-ccdf-4066-b3d8-246326d48806-kube-api-access-nz77s\") pod \"redhat-marketplace-lnckw\" (UID: \"da7c493c-ccdf-4066-b3d8-246326d48806\") " pod="openshift-marketplace/redhat-marketplace-lnckw" Dec 09 16:21:37 crc kubenswrapper[4894]: I1209 16:21:37.089905 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnckw" Dec 09 16:21:37 crc kubenswrapper[4894]: I1209 16:21:37.574497 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnckw"] Dec 09 16:21:37 crc kubenswrapper[4894]: I1209 16:21:37.790734 4894 generic.go:334] "Generic (PLEG): container finished" podID="da7c493c-ccdf-4066-b3d8-246326d48806" containerID="b4ccb120a4037ff09854c58a1db222e415b2286ea3e125545689b6dd086601c5" exitCode=0 Dec 09 16:21:37 crc kubenswrapper[4894]: I1209 16:21:37.790793 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnckw" event={"ID":"da7c493c-ccdf-4066-b3d8-246326d48806","Type":"ContainerDied","Data":"b4ccb120a4037ff09854c58a1db222e415b2286ea3e125545689b6dd086601c5"} Dec 09 16:21:37 crc kubenswrapper[4894]: I1209 16:21:37.790889 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnckw" event={"ID":"da7c493c-ccdf-4066-b3d8-246326d48806","Type":"ContainerStarted","Data":"20b5c02f811e97723cfd68fb662a1cf287629b4b429cdf9a5900273980508570"} Dec 09 16:21:38 crc kubenswrapper[4894]: I1209 16:21:38.803001 4894 generic.go:334] "Generic (PLEG): container finished" podID="da7c493c-ccdf-4066-b3d8-246326d48806" containerID="337f8a0ad408805b205c222b4ba9104f12427069fddd95a806c90a97f8548e58" exitCode=0 Dec 09 16:21:38 crc kubenswrapper[4894]: I1209 16:21:38.803088 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnckw" event={"ID":"da7c493c-ccdf-4066-b3d8-246326d48806","Type":"ContainerDied","Data":"337f8a0ad408805b205c222b4ba9104f12427069fddd95a806c90a97f8548e58"} Dec 09 16:21:39 crc kubenswrapper[4894]: I1209 16:21:39.812167 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnckw" event={"ID":"da7c493c-ccdf-4066-b3d8-246326d48806","Type":"ContainerStarted","Data":"13fdb5d4e2ac5ece60b25d11fdd7e15c28f590d5a34fe19dd5f143133b88da13"} Dec 09 16:21:39 crc kubenswrapper[4894]: I1209 16:21:39.835703 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lnckw" podStartSLOduration=2.312898181 podStartE2EDuration="3.835686695s" podCreationTimestamp="2025-12-09 16:21:36 +0000 UTC" firstStartedPulling="2025-12-09 16:21:37.792312839 +0000 UTC m=+2992.111523518" lastFinishedPulling="2025-12-09 16:21:39.315101363 +0000 UTC m=+2993.634312032" observedRunningTime="2025-12-09 16:21:39.829889648 +0000 UTC m=+2994.149100317" watchObservedRunningTime="2025-12-09 16:21:39.835686695 +0000 UTC m=+2994.154897364" Dec 09 16:21:44 crc kubenswrapper[4894]: I1209 16:21:44.106592 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:21:44 crc kubenswrapper[4894]: E1209 16:21:44.107134 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:21:47 crc kubenswrapper[4894]: I1209 16:21:47.090473 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lnckw" Dec 09 16:21:47 crc kubenswrapper[4894]: I1209 16:21:47.091077 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lnckw" Dec 09 16:21:47 crc kubenswrapper[4894]: I1209 16:21:47.146059 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lnckw" Dec 09 16:21:47 crc kubenswrapper[4894]: I1209 16:21:47.934421 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lnckw" Dec 09 16:21:47 crc kubenswrapper[4894]: I1209 16:21:47.988153 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnckw"] Dec 09 16:21:49 crc kubenswrapper[4894]: I1209 16:21:49.910072 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lnckw" podUID="da7c493c-ccdf-4066-b3d8-246326d48806" containerName="registry-server" containerID="cri-o://13fdb5d4e2ac5ece60b25d11fdd7e15c28f590d5a34fe19dd5f143133b88da13" gracePeriod=2 Dec 09 16:21:50 crc kubenswrapper[4894]: I1209 16:21:50.377710 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnckw" Dec 09 16:21:50 crc kubenswrapper[4894]: I1209 16:21:50.444631 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da7c493c-ccdf-4066-b3d8-246326d48806-utilities\") pod \"da7c493c-ccdf-4066-b3d8-246326d48806\" (UID: \"da7c493c-ccdf-4066-b3d8-246326d48806\") " Dec 09 16:21:50 crc kubenswrapper[4894]: I1209 16:21:50.444728 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da7c493c-ccdf-4066-b3d8-246326d48806-catalog-content\") pod \"da7c493c-ccdf-4066-b3d8-246326d48806\" (UID: \"da7c493c-ccdf-4066-b3d8-246326d48806\") " Dec 09 16:21:50 crc kubenswrapper[4894]: I1209 16:21:50.444758 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nz77s\" (UniqueName: \"kubernetes.io/projected/da7c493c-ccdf-4066-b3d8-246326d48806-kube-api-access-nz77s\") pod \"da7c493c-ccdf-4066-b3d8-246326d48806\" (UID: \"da7c493c-ccdf-4066-b3d8-246326d48806\") " Dec 09 16:21:50 crc kubenswrapper[4894]: I1209 16:21:50.445496 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da7c493c-ccdf-4066-b3d8-246326d48806-utilities" (OuterVolumeSpecName: "utilities") pod "da7c493c-ccdf-4066-b3d8-246326d48806" (UID: "da7c493c-ccdf-4066-b3d8-246326d48806"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:21:50 crc kubenswrapper[4894]: I1209 16:21:50.450127 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da7c493c-ccdf-4066-b3d8-246326d48806-kube-api-access-nz77s" (OuterVolumeSpecName: "kube-api-access-nz77s") pod "da7c493c-ccdf-4066-b3d8-246326d48806" (UID: "da7c493c-ccdf-4066-b3d8-246326d48806"). InnerVolumeSpecName "kube-api-access-nz77s". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:21:50 crc kubenswrapper[4894]: I1209 16:21:50.462846 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/da7c493c-ccdf-4066-b3d8-246326d48806-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "da7c493c-ccdf-4066-b3d8-246326d48806" (UID: "da7c493c-ccdf-4066-b3d8-246326d48806"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:21:50 crc kubenswrapper[4894]: I1209 16:21:50.545898 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da7c493c-ccdf-4066-b3d8-246326d48806-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:21:50 crc kubenswrapper[4894]: I1209 16:21:50.545936 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nz77s\" (UniqueName: \"kubernetes.io/projected/da7c493c-ccdf-4066-b3d8-246326d48806-kube-api-access-nz77s\") on node \"crc\" DevicePath \"\"" Dec 09 16:21:50 crc kubenswrapper[4894]: I1209 16:21:50.545952 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da7c493c-ccdf-4066-b3d8-246326d48806-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:21:50 crc kubenswrapper[4894]: I1209 16:21:50.925484 4894 generic.go:334] "Generic (PLEG): container finished" podID="da7c493c-ccdf-4066-b3d8-246326d48806" containerID="13fdb5d4e2ac5ece60b25d11fdd7e15c28f590d5a34fe19dd5f143133b88da13" exitCode=0 Dec 09 16:21:50 crc kubenswrapper[4894]: I1209 16:21:50.925582 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lnckw" Dec 09 16:21:50 crc kubenswrapper[4894]: I1209 16:21:50.927277 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnckw" event={"ID":"da7c493c-ccdf-4066-b3d8-246326d48806","Type":"ContainerDied","Data":"13fdb5d4e2ac5ece60b25d11fdd7e15c28f590d5a34fe19dd5f143133b88da13"} Dec 09 16:21:50 crc kubenswrapper[4894]: I1209 16:21:50.927597 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lnckw" event={"ID":"da7c493c-ccdf-4066-b3d8-246326d48806","Type":"ContainerDied","Data":"20b5c02f811e97723cfd68fb662a1cf287629b4b429cdf9a5900273980508570"} Dec 09 16:21:50 crc kubenswrapper[4894]: I1209 16:21:50.927690 4894 scope.go:117] "RemoveContainer" containerID="13fdb5d4e2ac5ece60b25d11fdd7e15c28f590d5a34fe19dd5f143133b88da13" Dec 09 16:21:50 crc kubenswrapper[4894]: I1209 16:21:50.955182 4894 scope.go:117] "RemoveContainer" containerID="337f8a0ad408805b205c222b4ba9104f12427069fddd95a806c90a97f8548e58" Dec 09 16:21:50 crc kubenswrapper[4894]: I1209 16:21:50.990938 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnckw"] Dec 09 16:21:51 crc kubenswrapper[4894]: I1209 16:21:51.001767 4894 scope.go:117] "RemoveContainer" containerID="b4ccb120a4037ff09854c58a1db222e415b2286ea3e125545689b6dd086601c5" Dec 09 16:21:51 crc kubenswrapper[4894]: I1209 16:21:51.011616 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lnckw"] Dec 09 16:21:51 crc kubenswrapper[4894]: I1209 16:21:51.043809 4894 scope.go:117] "RemoveContainer" containerID="13fdb5d4e2ac5ece60b25d11fdd7e15c28f590d5a34fe19dd5f143133b88da13" Dec 09 16:21:51 crc kubenswrapper[4894]: E1209 16:21:51.044453 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13fdb5d4e2ac5ece60b25d11fdd7e15c28f590d5a34fe19dd5f143133b88da13\": container with ID starting with 13fdb5d4e2ac5ece60b25d11fdd7e15c28f590d5a34fe19dd5f143133b88da13 not found: ID does not exist" containerID="13fdb5d4e2ac5ece60b25d11fdd7e15c28f590d5a34fe19dd5f143133b88da13" Dec 09 16:21:51 crc kubenswrapper[4894]: I1209 16:21:51.044506 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13fdb5d4e2ac5ece60b25d11fdd7e15c28f590d5a34fe19dd5f143133b88da13"} err="failed to get container status \"13fdb5d4e2ac5ece60b25d11fdd7e15c28f590d5a34fe19dd5f143133b88da13\": rpc error: code = NotFound desc = could not find container \"13fdb5d4e2ac5ece60b25d11fdd7e15c28f590d5a34fe19dd5f143133b88da13\": container with ID starting with 13fdb5d4e2ac5ece60b25d11fdd7e15c28f590d5a34fe19dd5f143133b88da13 not found: ID does not exist" Dec 09 16:21:51 crc kubenswrapper[4894]: I1209 16:21:51.044541 4894 scope.go:117] "RemoveContainer" containerID="337f8a0ad408805b205c222b4ba9104f12427069fddd95a806c90a97f8548e58" Dec 09 16:21:51 crc kubenswrapper[4894]: E1209 16:21:51.045551 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"337f8a0ad408805b205c222b4ba9104f12427069fddd95a806c90a97f8548e58\": container with ID starting with 337f8a0ad408805b205c222b4ba9104f12427069fddd95a806c90a97f8548e58 not found: ID does not exist" containerID="337f8a0ad408805b205c222b4ba9104f12427069fddd95a806c90a97f8548e58" Dec 09 16:21:51 crc kubenswrapper[4894]: I1209 16:21:51.045581 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"337f8a0ad408805b205c222b4ba9104f12427069fddd95a806c90a97f8548e58"} err="failed to get container status \"337f8a0ad408805b205c222b4ba9104f12427069fddd95a806c90a97f8548e58\": rpc error: code = NotFound desc = could not find container \"337f8a0ad408805b205c222b4ba9104f12427069fddd95a806c90a97f8548e58\": container with ID starting with 337f8a0ad408805b205c222b4ba9104f12427069fddd95a806c90a97f8548e58 not found: ID does not exist" Dec 09 16:21:51 crc kubenswrapper[4894]: I1209 16:21:51.045602 4894 scope.go:117] "RemoveContainer" containerID="b4ccb120a4037ff09854c58a1db222e415b2286ea3e125545689b6dd086601c5" Dec 09 16:21:51 crc kubenswrapper[4894]: E1209 16:21:51.046057 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b4ccb120a4037ff09854c58a1db222e415b2286ea3e125545689b6dd086601c5\": container with ID starting with b4ccb120a4037ff09854c58a1db222e415b2286ea3e125545689b6dd086601c5 not found: ID does not exist" containerID="b4ccb120a4037ff09854c58a1db222e415b2286ea3e125545689b6dd086601c5" Dec 09 16:21:51 crc kubenswrapper[4894]: I1209 16:21:51.046088 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b4ccb120a4037ff09854c58a1db222e415b2286ea3e125545689b6dd086601c5"} err="failed to get container status \"b4ccb120a4037ff09854c58a1db222e415b2286ea3e125545689b6dd086601c5\": rpc error: code = NotFound desc = could not find container \"b4ccb120a4037ff09854c58a1db222e415b2286ea3e125545689b6dd086601c5\": container with ID starting with b4ccb120a4037ff09854c58a1db222e415b2286ea3e125545689b6dd086601c5 not found: ID does not exist" Dec 09 16:21:51 crc kubenswrapper[4894]: I1209 16:21:51.941908 4894 generic.go:334] "Generic (PLEG): container finished" podID="9625f994-137d-4957-a9f9-40bf22bf832b" containerID="bf1ef71345d27d6192671dad28d741115ed4f1b805521d72820615af465a693b" exitCode=0 Dec 09 16:21:51 crc kubenswrapper[4894]: I1209 16:21:51.942276 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" event={"ID":"9625f994-137d-4957-a9f9-40bf22bf832b","Type":"ContainerDied","Data":"bf1ef71345d27d6192671dad28d741115ed4f1b805521d72820615af465a693b"} Dec 09 16:21:52 crc kubenswrapper[4894]: I1209 16:21:52.154597 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="da7c493c-ccdf-4066-b3d8-246326d48806" path="/var/lib/kubelet/pods/da7c493c-ccdf-4066-b3d8-246326d48806/volumes" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.383381 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.506294 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-migration-ssh-key-0\") pod \"9625f994-137d-4957-a9f9-40bf22bf832b\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.506375 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pm8nx\" (UniqueName: \"kubernetes.io/projected/9625f994-137d-4957-a9f9-40bf22bf832b-kube-api-access-pm8nx\") pod \"9625f994-137d-4957-a9f9-40bf22bf832b\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.506518 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-migration-ssh-key-1\") pod \"9625f994-137d-4957-a9f9-40bf22bf832b\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.506547 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9625f994-137d-4957-a9f9-40bf22bf832b-nova-extra-config-0\") pod \"9625f994-137d-4957-a9f9-40bf22bf832b\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.506588 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-ceph\") pod \"9625f994-137d-4957-a9f9-40bf22bf832b\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.506650 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-inventory\") pod \"9625f994-137d-4957-a9f9-40bf22bf832b\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.506676 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/9625f994-137d-4957-a9f9-40bf22bf832b-ceph-nova-0\") pod \"9625f994-137d-4957-a9f9-40bf22bf832b\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.506701 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-cell1-compute-config-1\") pod \"9625f994-137d-4957-a9f9-40bf22bf832b\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.506725 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-cell1-compute-config-0\") pod \"9625f994-137d-4957-a9f9-40bf22bf832b\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.506759 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-ssh-key\") pod \"9625f994-137d-4957-a9f9-40bf22bf832b\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.506779 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-custom-ceph-combined-ca-bundle\") pod \"9625f994-137d-4957-a9f9-40bf22bf832b\" (UID: \"9625f994-137d-4957-a9f9-40bf22bf832b\") " Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.512471 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-custom-ceph-combined-ca-bundle" (OuterVolumeSpecName: "nova-custom-ceph-combined-ca-bundle") pod "9625f994-137d-4957-a9f9-40bf22bf832b" (UID: "9625f994-137d-4957-a9f9-40bf22bf832b"). InnerVolumeSpecName "nova-custom-ceph-combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.512826 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-ceph" (OuterVolumeSpecName: "ceph") pod "9625f994-137d-4957-a9f9-40bf22bf832b" (UID: "9625f994-137d-4957-a9f9-40bf22bf832b"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.520149 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9625f994-137d-4957-a9f9-40bf22bf832b-kube-api-access-pm8nx" (OuterVolumeSpecName: "kube-api-access-pm8nx") pod "9625f994-137d-4957-a9f9-40bf22bf832b" (UID: "9625f994-137d-4957-a9f9-40bf22bf832b"). InnerVolumeSpecName "kube-api-access-pm8nx". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.542136 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-cell1-compute-config-1" (OuterVolumeSpecName: "nova-cell1-compute-config-1") pod "9625f994-137d-4957-a9f9-40bf22bf832b" (UID: "9625f994-137d-4957-a9f9-40bf22bf832b"). InnerVolumeSpecName "nova-cell1-compute-config-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.543548 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-migration-ssh-key-1" (OuterVolumeSpecName: "nova-migration-ssh-key-1") pod "9625f994-137d-4957-a9f9-40bf22bf832b" (UID: "9625f994-137d-4957-a9f9-40bf22bf832b"). InnerVolumeSpecName "nova-migration-ssh-key-1". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.545054 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-migration-ssh-key-0" (OuterVolumeSpecName: "nova-migration-ssh-key-0") pod "9625f994-137d-4957-a9f9-40bf22bf832b" (UID: "9625f994-137d-4957-a9f9-40bf22bf832b"). InnerVolumeSpecName "nova-migration-ssh-key-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.548941 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9625f994-137d-4957-a9f9-40bf22bf832b-ceph-nova-0" (OuterVolumeSpecName: "ceph-nova-0") pod "9625f994-137d-4957-a9f9-40bf22bf832b" (UID: "9625f994-137d-4957-a9f9-40bf22bf832b"). InnerVolumeSpecName "ceph-nova-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.550091 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9625f994-137d-4957-a9f9-40bf22bf832b-nova-extra-config-0" (OuterVolumeSpecName: "nova-extra-config-0") pod "9625f994-137d-4957-a9f9-40bf22bf832b" (UID: "9625f994-137d-4957-a9f9-40bf22bf832b"). InnerVolumeSpecName "nova-extra-config-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.551547 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "9625f994-137d-4957-a9f9-40bf22bf832b" (UID: "9625f994-137d-4957-a9f9-40bf22bf832b"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.552843 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-cell1-compute-config-0" (OuterVolumeSpecName: "nova-cell1-compute-config-0") pod "9625f994-137d-4957-a9f9-40bf22bf832b" (UID: "9625f994-137d-4957-a9f9-40bf22bf832b"). InnerVolumeSpecName "nova-cell1-compute-config-0". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.565407 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-inventory" (OuterVolumeSpecName: "inventory") pod "9625f994-137d-4957-a9f9-40bf22bf832b" (UID: "9625f994-137d-4957-a9f9-40bf22bf832b"). InnerVolumeSpecName "inventory". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.608877 4894 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-0\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-migration-ssh-key-0\") on node \"crc\" DevicePath \"\"" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.608915 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pm8nx\" (UniqueName: \"kubernetes.io/projected/9625f994-137d-4957-a9f9-40bf22bf832b-kube-api-access-pm8nx\") on node \"crc\" DevicePath \"\"" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.608929 4894 reconciler_common.go:293] "Volume detached for volume \"nova-migration-ssh-key-1\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-migration-ssh-key-1\") on node \"crc\" DevicePath \"\"" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.608944 4894 reconciler_common.go:293] "Volume detached for volume \"nova-extra-config-0\" (UniqueName: \"kubernetes.io/configmap/9625f994-137d-4957-a9f9-40bf22bf832b-nova-extra-config-0\") on node \"crc\" DevicePath \"\"" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.608956 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.608967 4894 reconciler_common.go:293] "Volume detached for volume \"inventory\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-inventory\") on node \"crc\" DevicePath \"\"" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.608977 4894 reconciler_common.go:293] "Volume detached for volume \"ceph-nova-0\" (UniqueName: \"kubernetes.io/configmap/9625f994-137d-4957-a9f9-40bf22bf832b-ceph-nova-0\") on node \"crc\" DevicePath \"\"" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.608989 4894 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-1\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-cell1-compute-config-1\") on node \"crc\" DevicePath \"\"" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.609000 4894 reconciler_common.go:293] "Volume detached for volume \"nova-cell1-compute-config-0\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-cell1-compute-config-0\") on node \"crc\" DevicePath \"\"" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.609010 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.609021 4894 reconciler_common.go:293] "Volume detached for volume \"nova-custom-ceph-combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9625f994-137d-4957-a9f9-40bf22bf832b-nova-custom-ceph-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.964543 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" event={"ID":"9625f994-137d-4957-a9f9-40bf22bf832b","Type":"ContainerDied","Data":"523f77a99e598917506b6d282f93c032ea65adbf4e2f71edd6836d8ab10b5bf0"} Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.966166 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="523f77a99e598917506b6d282f93c032ea65adbf4e2f71edd6836d8ab10b5bf0" Dec 09 16:21:53 crc kubenswrapper[4894]: I1209 16:21:53.964612 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29" Dec 09 16:21:57 crc kubenswrapper[4894]: I1209 16:21:57.106625 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:21:57 crc kubenswrapper[4894]: E1209 16:21:57.107743 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.658325 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 09 16:22:08 crc kubenswrapper[4894]: E1209 16:22:08.659166 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da7c493c-ccdf-4066-b3d8-246326d48806" containerName="extract-content" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.659178 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="da7c493c-ccdf-4066-b3d8-246326d48806" containerName="extract-content" Dec 09 16:22:08 crc kubenswrapper[4894]: E1209 16:22:08.659187 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da7c493c-ccdf-4066-b3d8-246326d48806" containerName="extract-utilities" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.659193 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="da7c493c-ccdf-4066-b3d8-246326d48806" containerName="extract-utilities" Dec 09 16:22:08 crc kubenswrapper[4894]: E1209 16:22:08.659214 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da7c493c-ccdf-4066-b3d8-246326d48806" containerName="registry-server" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.659220 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="da7c493c-ccdf-4066-b3d8-246326d48806" containerName="registry-server" Dec 09 16:22:08 crc kubenswrapper[4894]: E1209 16:22:08.659232 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9625f994-137d-4957-a9f9-40bf22bf832b" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.659240 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="9625f994-137d-4957-a9f9-40bf22bf832b" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.659407 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="9625f994-137d-4957-a9f9-40bf22bf832b" containerName="nova-custom-ceph-edpm-deployment-openstack-edpm-ipam" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.659418 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="da7c493c-ccdf-4066-b3d8-246326d48806" containerName="registry-server" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.660310 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.662651 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceph-conf-files" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.662838 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-volume-volume1-config-data" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.677093 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.696956 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73446a0b-19de-46e4-ba91-d183921ff32e-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.697007 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.697032 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-sys\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.697051 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.697066 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.697085 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-dev\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.697110 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.697135 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/73446a0b-19de-46e4-ba91-d183921ff32e-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.697160 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rf7m\" (UniqueName: \"kubernetes.io/projected/73446a0b-19de-46e4-ba91-d183921ff32e-kube-api-access-5rf7m\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.697193 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.697215 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73446a0b-19de-46e4-ba91-d183921ff32e-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.697237 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.697255 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.697287 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-run\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.697321 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73446a0b-19de-46e4-ba91-d183921ff32e-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.697374 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73446a0b-19de-46e4-ba91-d183921ff32e-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.702861 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-backup-0"] Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.704433 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.706359 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-backup-config-data" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.730942 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.798677 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.798725 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c22d31c-b595-4f8b-bded-c20a085ab1fd-scripts\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.798757 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-run\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.798808 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-dev\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.798947 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-etc-machine-id\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.799025 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-run\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.799066 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-lib-modules\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.799087 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73446a0b-19de-46e4-ba91-d183921ff32e-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.799107 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c22d31c-b595-4f8b-bded-c20a085ab1fd-config-data\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.799130 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.799348 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73446a0b-19de-46e4-ba91-d183921ff32e-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.799433 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-run\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.799484 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-sys\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.799517 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73446a0b-19de-46e4-ba91-d183921ff32e-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.799601 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.799673 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.799747 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-sys\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.799840 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.800215 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.799916 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-sys\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.800157 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-var-locks-brick\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.799950 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-var-lib-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.800384 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0c22d31c-b595-4f8b-bded-c20a085ab1fd-ceph\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.800423 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-dev\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.800464 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.800505 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.800505 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-dev\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.800552 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.800578 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.800651 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/73446a0b-19de-46e4-ba91-d183921ff32e-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.800682 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rf7m\" (UniqueName: \"kubernetes.io/projected/73446a0b-19de-46e4-ba91-d183921ff32e-kube-api-access-5rf7m\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.800750 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.800792 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73446a0b-19de-46e4-ba91-d183921ff32e-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.800834 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.800856 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.801045 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-etc-nvme\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.801086 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-lib-modules\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.800332 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-etc-iscsi\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.801131 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/73446a0b-19de-46e4-ba91-d183921ff32e-var-locks-cinder\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.804916 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/73446a0b-19de-46e4-ba91-d183921ff32e-config-data-custom\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.805863 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/73446a0b-19de-46e4-ba91-d183921ff32e-config-data\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.811941 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/73446a0b-19de-46e4-ba91-d183921ff32e-scripts\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.812825 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/73446a0b-19de-46e4-ba91-d183921ff32e-combined-ca-bundle\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.819333 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/73446a0b-19de-46e4-ba91-d183921ff32e-ceph\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.819977 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rf7m\" (UniqueName: \"kubernetes.io/projected/73446a0b-19de-46e4-ba91-d183921ff32e-kube-api-access-5rf7m\") pod \"cinder-volume-volume1-0\" (UID: \"73446a0b-19de-46e4-ba91-d183921ff32e\") " pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.901518 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c22d31c-b595-4f8b-bded-c20a085ab1fd-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.901585 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.901613 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c22d31c-b595-4f8b-bded-c20a085ab1fd-scripts\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.901810 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-dev\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.901881 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-lib-modules\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.901888 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-dev\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.901919 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c22d31c-b595-4f8b-bded-c20a085ab1fd-config-data\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.901945 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wlkrg\" (UniqueName: \"kubernetes.io/projected/0c22d31c-b595-4f8b-bded-c20a085ab1fd-kube-api-access-wlkrg\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.901951 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-lib-modules\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.901980 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.902054 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-cinder\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-var-locks-cinder\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.902140 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-run\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.902160 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-sys\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.902192 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-run\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.902212 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.902232 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-sys\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.902207 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-etc-iscsi\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.902251 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-var-locks-brick\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.902245 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c22d31c-b595-4f8b-bded-c20a085ab1fd-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.902356 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0c22d31c-b595-4f8b-bded-c20a085ab1fd-ceph\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.902384 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.902428 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.902459 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.902474 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-cinder\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-var-lib-cinder\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.902618 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-etc-machine-id\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.902664 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/0c22d31c-b595-4f8b-bded-c20a085ab1fd-etc-nvme\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.904605 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/0c22d31c-b595-4f8b-bded-c20a085ab1fd-scripts\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.905244 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0c22d31c-b595-4f8b-bded-c20a085ab1fd-config-data\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.905496 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/0c22d31c-b595-4f8b-bded-c20a085ab1fd-ceph\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:08 crc kubenswrapper[4894]: I1209 16:22:08.980182 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.003357 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c22d31c-b595-4f8b-bded-c20a085ab1fd-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.003439 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c22d31c-b595-4f8b-bded-c20a085ab1fd-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.003526 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wlkrg\" (UniqueName: \"kubernetes.io/projected/0c22d31c-b595-4f8b-bded-c20a085ab1fd-kube-api-access-wlkrg\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.009233 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/0c22d31c-b595-4f8b-bded-c20a085ab1fd-config-data-custom\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.009530 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0c22d31c-b595-4f8b-bded-c20a085ab1fd-combined-ca-bundle\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.023926 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wlkrg\" (UniqueName: \"kubernetes.io/projected/0c22d31c-b595-4f8b-bded-c20a085ab1fd-kube-api-access-wlkrg\") pod \"cinder-backup-0\" (UID: \"0c22d31c-b595-4f8b-bded-c20a085ab1fd\") " pod="openstack/cinder-backup-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.024336 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-backup-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.243269 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-create-rn6kt"] Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.245073 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-rn6kt" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.253128 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-rn6kt"] Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.316918 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3e1b2cf-f96b-450b-86be-f935ae75d0f0-operator-scripts\") pod \"manila-db-create-rn6kt\" (UID: \"a3e1b2cf-f96b-450b-86be-f935ae75d0f0\") " pod="openstack/manila-db-create-rn6kt" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.316965 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cjrgz\" (UniqueName: \"kubernetes.io/projected/a3e1b2cf-f96b-450b-86be-f935ae75d0f0-kube-api-access-cjrgz\") pod \"manila-db-create-rn6kt\" (UID: \"a3e1b2cf-f96b-450b-86be-f935ae75d0f0\") " pod="openstack/manila-db-create-rn6kt" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.335311 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-bf42-account-create-update-6qjdg"] Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.336371 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-bf42-account-create-update-6qjdg" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.338530 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-db-secret" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.344384 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-bf42-account-create-update-6qjdg"] Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.419404 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zw65z\" (UniqueName: \"kubernetes.io/projected/fa0ee67f-8fa7-42e0-a524-0bf863f1bc67-kube-api-access-zw65z\") pod \"manila-bf42-account-create-update-6qjdg\" (UID: \"fa0ee67f-8fa7-42e0-a524-0bf863f1bc67\") " pod="openstack/manila-bf42-account-create-update-6qjdg" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.419470 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3e1b2cf-f96b-450b-86be-f935ae75d0f0-operator-scripts\") pod \"manila-db-create-rn6kt\" (UID: \"a3e1b2cf-f96b-450b-86be-f935ae75d0f0\") " pod="openstack/manila-db-create-rn6kt" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.419495 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cjrgz\" (UniqueName: \"kubernetes.io/projected/a3e1b2cf-f96b-450b-86be-f935ae75d0f0-kube-api-access-cjrgz\") pod \"manila-db-create-rn6kt\" (UID: \"a3e1b2cf-f96b-450b-86be-f935ae75d0f0\") " pod="openstack/manila-db-create-rn6kt" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.419564 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa0ee67f-8fa7-42e0-a524-0bf863f1bc67-operator-scripts\") pod \"manila-bf42-account-create-update-6qjdg\" (UID: \"fa0ee67f-8fa7-42e0-a524-0bf863f1bc67\") " pod="openstack/manila-bf42-account-create-update-6qjdg" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.420341 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3e1b2cf-f96b-450b-86be-f935ae75d0f0-operator-scripts\") pod \"manila-db-create-rn6kt\" (UID: \"a3e1b2cf-f96b-450b-86be-f935ae75d0f0\") " pod="openstack/manila-db-create-rn6kt" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.438709 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cjrgz\" (UniqueName: \"kubernetes.io/projected/a3e1b2cf-f96b-450b-86be-f935ae75d0f0-kube-api-access-cjrgz\") pod \"manila-db-create-rn6kt\" (UID: \"a3e1b2cf-f96b-450b-86be-f935ae75d0f0\") " pod="openstack/manila-db-create-rn6kt" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.520236 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa0ee67f-8fa7-42e0-a524-0bf863f1bc67-operator-scripts\") pod \"manila-bf42-account-create-update-6qjdg\" (UID: \"fa0ee67f-8fa7-42e0-a524-0bf863f1bc67\") " pod="openstack/manila-bf42-account-create-update-6qjdg" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.520586 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zw65z\" (UniqueName: \"kubernetes.io/projected/fa0ee67f-8fa7-42e0-a524-0bf863f1bc67-kube-api-access-zw65z\") pod \"manila-bf42-account-create-update-6qjdg\" (UID: \"fa0ee67f-8fa7-42e0-a524-0bf863f1bc67\") " pod="openstack/manila-bf42-account-create-update-6qjdg" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.521823 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa0ee67f-8fa7-42e0-a524-0bf863f1bc67-operator-scripts\") pod \"manila-bf42-account-create-update-6qjdg\" (UID: \"fa0ee67f-8fa7-42e0-a524-0bf863f1bc67\") " pod="openstack/manila-bf42-account-create-update-6qjdg" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.542357 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.543621 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.543748 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.572041 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.572196 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.572849 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-pj76v" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.573051 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.573046 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zw65z\" (UniqueName: \"kubernetes.io/projected/fa0ee67f-8fa7-42e0-a524-0bf863f1bc67-kube-api-access-zw65z\") pod \"manila-bf42-account-create-update-6qjdg\" (UID: \"fa0ee67f-8fa7-42e0-a524-0bf863f1bc67\") " pod="openstack/manila-bf42-account-create-update-6qjdg" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.578741 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-rn6kt" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.641899 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.646456 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.649002 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.649214 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.663120 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-bf42-account-create-update-6qjdg" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.703391 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.718839 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-volume-volume1-0"] Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.735837 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/850b00b0-204b-4af9-a22d-f4264d3cb44f-scripts\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.736204 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/850b00b0-204b-4af9-a22d-f4264d3cb44f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.736226 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/850b00b0-204b-4af9-a22d-f4264d3cb44f-logs\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.736341 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2b58t\" (UniqueName: \"kubernetes.io/projected/850b00b0-204b-4af9-a22d-f4264d3cb44f-kube-api-access-2b58t\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.736453 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/850b00b0-204b-4af9-a22d-f4264d3cb44f-ceph\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.736483 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/850b00b0-204b-4af9-a22d-f4264d3cb44f-config-data\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.736517 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/850b00b0-204b-4af9-a22d-f4264d3cb44f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.736537 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/850b00b0-204b-4af9-a22d-f4264d3cb44f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.736596 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.772577 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-backup-0"] Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.838849 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09510dc1-caeb-42f3-ac0f-150c7ab8365c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.838913 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/850b00b0-204b-4af9-a22d-f4264d3cb44f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.838934 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/850b00b0-204b-4af9-a22d-f4264d3cb44f-logs\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.838952 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09510dc1-caeb-42f3-ac0f-150c7ab8365c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.838986 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09510dc1-caeb-42f3-ac0f-150c7ab8365c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.839005 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2b58t\" (UniqueName: \"kubernetes.io/projected/850b00b0-204b-4af9-a22d-f4264d3cb44f-kube-api-access-2b58t\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.839028 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/09510dc1-caeb-42f3-ac0f-150c7ab8365c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.839054 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/09510dc1-caeb-42f3-ac0f-150c7ab8365c-ceph\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.839081 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/850b00b0-204b-4af9-a22d-f4264d3cb44f-ceph\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.839100 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/850b00b0-204b-4af9-a22d-f4264d3cb44f-config-data\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.839124 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/850b00b0-204b-4af9-a22d-f4264d3cb44f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.839140 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/850b00b0-204b-4af9-a22d-f4264d3cb44f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.839158 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.839172 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/09510dc1-caeb-42f3-ac0f-150c7ab8365c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.839221 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/850b00b0-204b-4af9-a22d-f4264d3cb44f-scripts\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.839237 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09510dc1-caeb-42f3-ac0f-150c7ab8365c-logs\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.839255 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.839271 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6l9h9\" (UniqueName: \"kubernetes.io/projected/09510dc1-caeb-42f3-ac0f-150c7ab8365c-kube-api-access-6l9h9\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.839693 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/850b00b0-204b-4af9-a22d-f4264d3cb44f-logs\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.840462 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.840487 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/850b00b0-204b-4af9-a22d-f4264d3cb44f-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.847271 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/850b00b0-204b-4af9-a22d-f4264d3cb44f-ceph\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.847835 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/850b00b0-204b-4af9-a22d-f4264d3cb44f-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.851251 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/850b00b0-204b-4af9-a22d-f4264d3cb44f-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.853609 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/850b00b0-204b-4af9-a22d-f4264d3cb44f-config-data\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.856477 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2b58t\" (UniqueName: \"kubernetes.io/projected/850b00b0-204b-4af9-a22d-f4264d3cb44f-kube-api-access-2b58t\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.860130 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/850b00b0-204b-4af9-a22d-f4264d3cb44f-scripts\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.875582 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"glance-default-external-api-0\" (UID: \"850b00b0-204b-4af9-a22d-f4264d3cb44f\") " pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.890373 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.940483 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09510dc1-caeb-42f3-ac0f-150c7ab8365c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.940560 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09510dc1-caeb-42f3-ac0f-150c7ab8365c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.940612 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09510dc1-caeb-42f3-ac0f-150c7ab8365c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.940799 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/09510dc1-caeb-42f3-ac0f-150c7ab8365c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.940842 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/09510dc1-caeb-42f3-ac0f-150c7ab8365c-ceph\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.940906 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/09510dc1-caeb-42f3-ac0f-150c7ab8365c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.940984 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09510dc1-caeb-42f3-ac0f-150c7ab8365c-logs\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.941010 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.941032 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6l9h9\" (UniqueName: \"kubernetes.io/projected/09510dc1-caeb-42f3-ac0f-150c7ab8365c-kube-api-access-6l9h9\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.942933 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/09510dc1-caeb-42f3-ac0f-150c7ab8365c-logs\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.943127 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") device mount path \"/mnt/openstack/pv03\"" pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.946902 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/09510dc1-caeb-42f3-ac0f-150c7ab8365c-ceph\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.947205 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/09510dc1-caeb-42f3-ac0f-150c7ab8365c-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.948059 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/09510dc1-caeb-42f3-ac0f-150c7ab8365c-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.950191 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/09510dc1-caeb-42f3-ac0f-150c7ab8365c-scripts\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.953404 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/09510dc1-caeb-42f3-ac0f-150c7ab8365c-config-data\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.957584 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/09510dc1-caeb-42f3-ac0f-150c7ab8365c-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.982413 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6l9h9\" (UniqueName: \"kubernetes.io/projected/09510dc1-caeb-42f3-ac0f-150c7ab8365c-kube-api-access-6l9h9\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:09 crc kubenswrapper[4894]: I1209 16:22:09.984862 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"09510dc1-caeb-42f3-ac0f-150c7ab8365c\") " pod="openstack/glance-default-internal-api-0" Dec 09 16:22:10 crc kubenswrapper[4894]: I1209 16:22:10.144852 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"0c22d31c-b595-4f8b-bded-c20a085ab1fd","Type":"ContainerStarted","Data":"202f3a3b1798a6a8aec299ae2fe3617d6050d70bbf2cc9b955303766b23a80ef"} Dec 09 16:22:10 crc kubenswrapper[4894]: I1209 16:22:10.146324 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"73446a0b-19de-46e4-ba91-d183921ff32e","Type":"ContainerStarted","Data":"e09001d733eb1ac1b47d00df7591e7f31baeff3ec5ef6af521a72a25dd3ab6a5"} Dec 09 16:22:10 crc kubenswrapper[4894]: I1209 16:22:10.176519 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-create-rn6kt"] Dec 09 16:22:10 crc kubenswrapper[4894]: I1209 16:22:10.268142 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Dec 09 16:22:10 crc kubenswrapper[4894]: I1209 16:22:10.289121 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-bf42-account-create-update-6qjdg"] Dec 09 16:22:11 crc kubenswrapper[4894]: I1209 16:22:10.629207 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Dec 09 16:22:11 crc kubenswrapper[4894]: W1209 16:22:10.828322 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod850b00b0_204b_4af9_a22d_f4264d3cb44f.slice/crio-c5de1b73d95c8dcc64a1790320455d20527c18e867b98eafd779332872358aec WatchSource:0}: Error finding container c5de1b73d95c8dcc64a1790320455d20527c18e867b98eafd779332872358aec: Status 404 returned error can't find the container with id c5de1b73d95c8dcc64a1790320455d20527c18e867b98eafd779332872358aec Dec 09 16:22:11 crc kubenswrapper[4894]: I1209 16:22:10.912248 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Dec 09 16:22:11 crc kubenswrapper[4894]: W1209 16:22:10.951529 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod09510dc1_caeb_42f3_ac0f_150c7ab8365c.slice/crio-98a11a4f3c82c054a736b2a54ee5ad9e020e9aa2841ff7f2bdebcb202a7439ed WatchSource:0}: Error finding container 98a11a4f3c82c054a736b2a54ee5ad9e020e9aa2841ff7f2bdebcb202a7439ed: Status 404 returned error can't find the container with id 98a11a4f3c82c054a736b2a54ee5ad9e020e9aa2841ff7f2bdebcb202a7439ed Dec 09 16:22:11 crc kubenswrapper[4894]: I1209 16:22:11.106345 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:22:11 crc kubenswrapper[4894]: E1209 16:22:11.106903 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:22:11 crc kubenswrapper[4894]: I1209 16:22:11.173215 4894 generic.go:334] "Generic (PLEG): container finished" podID="a3e1b2cf-f96b-450b-86be-f935ae75d0f0" containerID="d48676851ecbcfe69d01635612289e472766580df9d455098ee694b7e2973346" exitCode=0 Dec 09 16:22:11 crc kubenswrapper[4894]: I1209 16:22:11.173742 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-rn6kt" event={"ID":"a3e1b2cf-f96b-450b-86be-f935ae75d0f0","Type":"ContainerDied","Data":"d48676851ecbcfe69d01635612289e472766580df9d455098ee694b7e2973346"} Dec 09 16:22:11 crc kubenswrapper[4894]: I1209 16:22:11.173806 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-rn6kt" event={"ID":"a3e1b2cf-f96b-450b-86be-f935ae75d0f0","Type":"ContainerStarted","Data":"983f702e3d64e67162936381940395bac18864e306de7a3bf30b8d9473dbe194"} Dec 09 16:22:11 crc kubenswrapper[4894]: I1209 16:22:11.176000 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"09510dc1-caeb-42f3-ac0f-150c7ab8365c","Type":"ContainerStarted","Data":"98a11a4f3c82c054a736b2a54ee5ad9e020e9aa2841ff7f2bdebcb202a7439ed"} Dec 09 16:22:11 crc kubenswrapper[4894]: I1209 16:22:11.178552 4894 generic.go:334] "Generic (PLEG): container finished" podID="fa0ee67f-8fa7-42e0-a524-0bf863f1bc67" containerID="a628f9d9624018f1fb0884d8760be48fb582b7fab403659d4ccee6b962734e9a" exitCode=0 Dec 09 16:22:11 crc kubenswrapper[4894]: I1209 16:22:11.178710 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-bf42-account-create-update-6qjdg" event={"ID":"fa0ee67f-8fa7-42e0-a524-0bf863f1bc67","Type":"ContainerDied","Data":"a628f9d9624018f1fb0884d8760be48fb582b7fab403659d4ccee6b962734e9a"} Dec 09 16:22:11 crc kubenswrapper[4894]: I1209 16:22:11.178742 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-bf42-account-create-update-6qjdg" event={"ID":"fa0ee67f-8fa7-42e0-a524-0bf863f1bc67","Type":"ContainerStarted","Data":"0c40902445978d81e5a827c67769b5cfbf073a7c0c61449630ad864ceabf3742"} Dec 09 16:22:11 crc kubenswrapper[4894]: I1209 16:22:11.181653 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"850b00b0-204b-4af9-a22d-f4264d3cb44f","Type":"ContainerStarted","Data":"c5de1b73d95c8dcc64a1790320455d20527c18e867b98eafd779332872358aec"} Dec 09 16:22:12 crc kubenswrapper[4894]: I1209 16:22:12.194673 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"850b00b0-204b-4af9-a22d-f4264d3cb44f","Type":"ContainerStarted","Data":"5632fba407676ad15aeb87356a03d27d37512a8478a541148cacc44e87d3cbf8"} Dec 09 16:22:12 crc kubenswrapper[4894]: I1209 16:22:12.197871 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"73446a0b-19de-46e4-ba91-d183921ff32e","Type":"ContainerStarted","Data":"1c5db914be04b348bd64d48b8b5227091358c8738f84b3344bb8c98c3e42c1ea"} Dec 09 16:22:12 crc kubenswrapper[4894]: I1209 16:22:12.197925 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-volume-volume1-0" event={"ID":"73446a0b-19de-46e4-ba91-d183921ff32e","Type":"ContainerStarted","Data":"ee80141c11b3f79fc4725b12dc438cd8e7d86bf36348737572e17debaa8c7d3e"} Dec 09 16:22:12 crc kubenswrapper[4894]: I1209 16:22:12.210151 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"09510dc1-caeb-42f3-ac0f-150c7ab8365c","Type":"ContainerStarted","Data":"a6e0e56ba3d1a56c9d8ecf476c1acca87c39281ec3da00dce48bc8c9bd77916f"} Dec 09 16:22:12 crc kubenswrapper[4894]: I1209 16:22:12.240506 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"0c22d31c-b595-4f8b-bded-c20a085ab1fd","Type":"ContainerStarted","Data":"e374170e36e18d05b6335f8269d51ed1096c092edfde6f34a8fa786ccb2c624e"} Dec 09 16:22:12 crc kubenswrapper[4894]: I1209 16:22:12.240555 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-backup-0" event={"ID":"0c22d31c-b595-4f8b-bded-c20a085ab1fd","Type":"ContainerStarted","Data":"712c576269c32efd818143ce5835684bfd97608f5b5bd164176f98c97ac00016"} Dec 09 16:22:12 crc kubenswrapper[4894]: I1209 16:22:12.245451 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-volume-volume1-0" podStartSLOduration=2.930697867 podStartE2EDuration="4.245426831s" podCreationTimestamp="2025-12-09 16:22:08 +0000 UTC" firstStartedPulling="2025-12-09 16:22:09.712934482 +0000 UTC m=+3024.032145151" lastFinishedPulling="2025-12-09 16:22:11.027663446 +0000 UTC m=+3025.346874115" observedRunningTime="2025-12-09 16:22:12.225590885 +0000 UTC m=+3026.544801574" watchObservedRunningTime="2025-12-09 16:22:12.245426831 +0000 UTC m=+3026.564637530" Dec 09 16:22:12 crc kubenswrapper[4894]: I1209 16:22:12.272361 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-backup-0" podStartSLOduration=3.029041374 podStartE2EDuration="4.272338638s" podCreationTimestamp="2025-12-09 16:22:08 +0000 UTC" firstStartedPulling="2025-12-09 16:22:09.783571551 +0000 UTC m=+3024.102782220" lastFinishedPulling="2025-12-09 16:22:11.026868815 +0000 UTC m=+3025.346079484" observedRunningTime="2025-12-09 16:22:12.269133121 +0000 UTC m=+3026.588343790" watchObservedRunningTime="2025-12-09 16:22:12.272338638 +0000 UTC m=+3026.591549307" Dec 09 16:22:12 crc kubenswrapper[4894]: I1209 16:22:12.657454 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-bf42-account-create-update-6qjdg" Dec 09 16:22:12 crc kubenswrapper[4894]: I1209 16:22:12.676327 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-rn6kt" Dec 09 16:22:12 crc kubenswrapper[4894]: I1209 16:22:12.829939 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3e1b2cf-f96b-450b-86be-f935ae75d0f0-operator-scripts\") pod \"a3e1b2cf-f96b-450b-86be-f935ae75d0f0\" (UID: \"a3e1b2cf-f96b-450b-86be-f935ae75d0f0\") " Dec 09 16:22:12 crc kubenswrapper[4894]: I1209 16:22:12.830093 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cjrgz\" (UniqueName: \"kubernetes.io/projected/a3e1b2cf-f96b-450b-86be-f935ae75d0f0-kube-api-access-cjrgz\") pod \"a3e1b2cf-f96b-450b-86be-f935ae75d0f0\" (UID: \"a3e1b2cf-f96b-450b-86be-f935ae75d0f0\") " Dec 09 16:22:12 crc kubenswrapper[4894]: I1209 16:22:12.830116 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa0ee67f-8fa7-42e0-a524-0bf863f1bc67-operator-scripts\") pod \"fa0ee67f-8fa7-42e0-a524-0bf863f1bc67\" (UID: \"fa0ee67f-8fa7-42e0-a524-0bf863f1bc67\") " Dec 09 16:22:12 crc kubenswrapper[4894]: I1209 16:22:12.830228 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zw65z\" (UniqueName: \"kubernetes.io/projected/fa0ee67f-8fa7-42e0-a524-0bf863f1bc67-kube-api-access-zw65z\") pod \"fa0ee67f-8fa7-42e0-a524-0bf863f1bc67\" (UID: \"fa0ee67f-8fa7-42e0-a524-0bf863f1bc67\") " Dec 09 16:22:12 crc kubenswrapper[4894]: I1209 16:22:12.830605 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa0ee67f-8fa7-42e0-a524-0bf863f1bc67-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "fa0ee67f-8fa7-42e0-a524-0bf863f1bc67" (UID: "fa0ee67f-8fa7-42e0-a524-0bf863f1bc67"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:22:12 crc kubenswrapper[4894]: I1209 16:22:12.831154 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a3e1b2cf-f96b-450b-86be-f935ae75d0f0-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a3e1b2cf-f96b-450b-86be-f935ae75d0f0" (UID: "a3e1b2cf-f96b-450b-86be-f935ae75d0f0"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:22:12 crc kubenswrapper[4894]: I1209 16:22:12.835479 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a3e1b2cf-f96b-450b-86be-f935ae75d0f0-kube-api-access-cjrgz" (OuterVolumeSpecName: "kube-api-access-cjrgz") pod "a3e1b2cf-f96b-450b-86be-f935ae75d0f0" (UID: "a3e1b2cf-f96b-450b-86be-f935ae75d0f0"). InnerVolumeSpecName "kube-api-access-cjrgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:22:12 crc kubenswrapper[4894]: I1209 16:22:12.840967 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa0ee67f-8fa7-42e0-a524-0bf863f1bc67-kube-api-access-zw65z" (OuterVolumeSpecName: "kube-api-access-zw65z") pod "fa0ee67f-8fa7-42e0-a524-0bf863f1bc67" (UID: "fa0ee67f-8fa7-42e0-a524-0bf863f1bc67"). InnerVolumeSpecName "kube-api-access-zw65z". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:22:12 crc kubenswrapper[4894]: I1209 16:22:12.932166 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a3e1b2cf-f96b-450b-86be-f935ae75d0f0-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:12 crc kubenswrapper[4894]: I1209 16:22:12.932197 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cjrgz\" (UniqueName: \"kubernetes.io/projected/a3e1b2cf-f96b-450b-86be-f935ae75d0f0-kube-api-access-cjrgz\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:12 crc kubenswrapper[4894]: I1209 16:22:12.932208 4894 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/fa0ee67f-8fa7-42e0-a524-0bf863f1bc67-operator-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:12 crc kubenswrapper[4894]: I1209 16:22:12.932218 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zw65z\" (UniqueName: \"kubernetes.io/projected/fa0ee67f-8fa7-42e0-a524-0bf863f1bc67-kube-api-access-zw65z\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:13 crc kubenswrapper[4894]: I1209 16:22:13.249819 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"850b00b0-204b-4af9-a22d-f4264d3cb44f","Type":"ContainerStarted","Data":"d52f0316092831f8b64d7b018f8e2137a263d01cc376a01ff4c50ff167feeb88"} Dec 09 16:22:13 crc kubenswrapper[4894]: I1209 16:22:13.251684 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-create-rn6kt" Dec 09 16:22:13 crc kubenswrapper[4894]: I1209 16:22:13.252142 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-create-rn6kt" event={"ID":"a3e1b2cf-f96b-450b-86be-f935ae75d0f0","Type":"ContainerDied","Data":"983f702e3d64e67162936381940395bac18864e306de7a3bf30b8d9473dbe194"} Dec 09 16:22:13 crc kubenswrapper[4894]: I1209 16:22:13.252194 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="983f702e3d64e67162936381940395bac18864e306de7a3bf30b8d9473dbe194" Dec 09 16:22:13 crc kubenswrapper[4894]: I1209 16:22:13.253618 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"09510dc1-caeb-42f3-ac0f-150c7ab8365c","Type":"ContainerStarted","Data":"55e5b814a90efd3191039a5b3244a7db54d59b796318f7fb34d76fea1d83c29f"} Dec 09 16:22:13 crc kubenswrapper[4894]: I1209 16:22:13.255244 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-bf42-account-create-update-6qjdg" Dec 09 16:22:13 crc kubenswrapper[4894]: I1209 16:22:13.255307 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-bf42-account-create-update-6qjdg" event={"ID":"fa0ee67f-8fa7-42e0-a524-0bf863f1bc67","Type":"ContainerDied","Data":"0c40902445978d81e5a827c67769b5cfbf073a7c0c61449630ad864ceabf3742"} Dec 09 16:22:13 crc kubenswrapper[4894]: I1209 16:22:13.255325 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0c40902445978d81e5a827c67769b5cfbf073a7c0c61449630ad864ceabf3742" Dec 09 16:22:13 crc kubenswrapper[4894]: I1209 16:22:13.291555 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=5.291540039 podStartE2EDuration="5.291540039s" podCreationTimestamp="2025-12-09 16:22:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:22:13.287460618 +0000 UTC m=+3027.606671287" watchObservedRunningTime="2025-12-09 16:22:13.291540039 +0000 UTC m=+3027.610750708" Dec 09 16:22:13 crc kubenswrapper[4894]: I1209 16:22:13.328596 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=5.32857318 podStartE2EDuration="5.32857318s" podCreationTimestamp="2025-12-09 16:22:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:22:13.323244665 +0000 UTC m=+3027.642455334" watchObservedRunningTime="2025-12-09 16:22:13.32857318 +0000 UTC m=+3027.647783849" Dec 09 16:22:13 crc kubenswrapper[4894]: I1209 16:22:13.981571 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.024994 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-backup-0" Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.652297 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-db-sync-jvrwg"] Dec 09 16:22:14 crc kubenswrapper[4894]: E1209 16:22:14.652864 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa0ee67f-8fa7-42e0-a524-0bf863f1bc67" containerName="mariadb-account-create-update" Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.652895 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa0ee67f-8fa7-42e0-a524-0bf863f1bc67" containerName="mariadb-account-create-update" Dec 09 16:22:14 crc kubenswrapper[4894]: E1209 16:22:14.652915 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a3e1b2cf-f96b-450b-86be-f935ae75d0f0" containerName="mariadb-database-create" Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.652925 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a3e1b2cf-f96b-450b-86be-f935ae75d0f0" containerName="mariadb-database-create" Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.653157 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="a3e1b2cf-f96b-450b-86be-f935ae75d0f0" containerName="mariadb-database-create" Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.653183 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa0ee67f-8fa7-42e0-a524-0bf863f1bc67" containerName="mariadb-account-create-update" Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.654033 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-jvrwg" Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.660022 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-g2lbf" Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.660124 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.661901 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-jvrwg"] Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.773169 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-combined-ca-bundle\") pod \"manila-db-sync-jvrwg\" (UID: \"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a\") " pod="openstack/manila-db-sync-jvrwg" Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.773244 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-job-config-data\") pod \"manila-db-sync-jvrwg\" (UID: \"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a\") " pod="openstack/manila-db-sync-jvrwg" Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.773275 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-config-data\") pod \"manila-db-sync-jvrwg\" (UID: \"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a\") " pod="openstack/manila-db-sync-jvrwg" Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.773326 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bg5lm\" (UniqueName: \"kubernetes.io/projected/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-kube-api-access-bg5lm\") pod \"manila-db-sync-jvrwg\" (UID: \"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a\") " pod="openstack/manila-db-sync-jvrwg" Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.874661 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-combined-ca-bundle\") pod \"manila-db-sync-jvrwg\" (UID: \"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a\") " pod="openstack/manila-db-sync-jvrwg" Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.874718 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-job-config-data\") pod \"manila-db-sync-jvrwg\" (UID: \"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a\") " pod="openstack/manila-db-sync-jvrwg" Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.874740 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-config-data\") pod \"manila-db-sync-jvrwg\" (UID: \"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a\") " pod="openstack/manila-db-sync-jvrwg" Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.874774 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bg5lm\" (UniqueName: \"kubernetes.io/projected/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-kube-api-access-bg5lm\") pod \"manila-db-sync-jvrwg\" (UID: \"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a\") " pod="openstack/manila-db-sync-jvrwg" Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.881813 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-job-config-data\") pod \"manila-db-sync-jvrwg\" (UID: \"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a\") " pod="openstack/manila-db-sync-jvrwg" Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.883517 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-config-data\") pod \"manila-db-sync-jvrwg\" (UID: \"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a\") " pod="openstack/manila-db-sync-jvrwg" Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.884276 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-combined-ca-bundle\") pod \"manila-db-sync-jvrwg\" (UID: \"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a\") " pod="openstack/manila-db-sync-jvrwg" Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.907759 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bg5lm\" (UniqueName: \"kubernetes.io/projected/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-kube-api-access-bg5lm\") pod \"manila-db-sync-jvrwg\" (UID: \"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a\") " pod="openstack/manila-db-sync-jvrwg" Dec 09 16:22:14 crc kubenswrapper[4894]: I1209 16:22:14.982340 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-jvrwg" Dec 09 16:22:15 crc kubenswrapper[4894]: I1209 16:22:15.564988 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-db-sync-jvrwg"] Dec 09 16:22:16 crc kubenswrapper[4894]: I1209 16:22:16.288426 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-jvrwg" event={"ID":"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a","Type":"ContainerStarted","Data":"3af203d1ff32e25047dcaf61fc68496d893d59fcca2dc3ca0ad571e813179d81"} Dec 09 16:22:19 crc kubenswrapper[4894]: I1209 16:22:19.180898 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-volume-volume1-0" Dec 09 16:22:19 crc kubenswrapper[4894]: I1209 16:22:19.279740 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-backup-0" Dec 09 16:22:19 crc kubenswrapper[4894]: I1209 16:22:19.891983 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 09 16:22:19 crc kubenswrapper[4894]: I1209 16:22:19.892346 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Dec 09 16:22:19 crc kubenswrapper[4894]: I1209 16:22:19.929401 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 09 16:22:19 crc kubenswrapper[4894]: I1209 16:22:19.933387 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Dec 09 16:22:20 crc kubenswrapper[4894]: I1209 16:22:20.269050 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 09 16:22:20 crc kubenswrapper[4894]: I1209 16:22:20.269418 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Dec 09 16:22:20 crc kubenswrapper[4894]: I1209 16:22:20.303947 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 09 16:22:20 crc kubenswrapper[4894]: I1209 16:22:20.335717 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Dec 09 16:22:20 crc kubenswrapper[4894]: I1209 16:22:20.340450 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 09 16:22:20 crc kubenswrapper[4894]: I1209 16:22:20.340471 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 09 16:22:20 crc kubenswrapper[4894]: I1209 16:22:20.340481 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Dec 09 16:22:20 crc kubenswrapper[4894]: I1209 16:22:20.340491 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Dec 09 16:22:21 crc kubenswrapper[4894]: I1209 16:22:21.350361 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-jvrwg" event={"ID":"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a","Type":"ContainerStarted","Data":"2047006fc445345d89f058d9944e8a8d61a4d649ec8743ceee1c32242849733f"} Dec 09 16:22:21 crc kubenswrapper[4894]: I1209 16:22:21.385472 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-db-sync-jvrwg" podStartSLOduration=2.601854458 podStartE2EDuration="7.385456094s" podCreationTimestamp="2025-12-09 16:22:14 +0000 UTC" firstStartedPulling="2025-12-09 16:22:15.549875872 +0000 UTC m=+3029.869086541" lastFinishedPulling="2025-12-09 16:22:20.333477498 +0000 UTC m=+3034.652688177" observedRunningTime="2025-12-09 16:22:21.379194034 +0000 UTC m=+3035.698404693" watchObservedRunningTime="2025-12-09 16:22:21.385456094 +0000 UTC m=+3035.704666763" Dec 09 16:22:22 crc kubenswrapper[4894]: I1209 16:22:22.360491 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 16:22:22 crc kubenswrapper[4894]: I1209 16:22:22.361202 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 16:22:22 crc kubenswrapper[4894]: I1209 16:22:22.361020 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 16:22:22 crc kubenswrapper[4894]: I1209 16:22:22.361352 4894 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Dec 09 16:22:22 crc kubenswrapper[4894]: I1209 16:22:22.582174 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 09 16:22:22 crc kubenswrapper[4894]: I1209 16:22:22.582354 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Dec 09 16:22:22 crc kubenswrapper[4894]: I1209 16:22:22.588035 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 09 16:22:22 crc kubenswrapper[4894]: I1209 16:22:22.658345 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Dec 09 16:22:25 crc kubenswrapper[4894]: I1209 16:22:25.106868 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:22:25 crc kubenswrapper[4894]: E1209 16:22:25.107970 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:22:33 crc kubenswrapper[4894]: I1209 16:22:33.473462 4894 generic.go:334] "Generic (PLEG): container finished" podID="1f88f6fb-7da5-47f6-b0b5-a33af1b6230a" containerID="2047006fc445345d89f058d9944e8a8d61a4d649ec8743ceee1c32242849733f" exitCode=0 Dec 09 16:22:33 crc kubenswrapper[4894]: I1209 16:22:33.473621 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-jvrwg" event={"ID":"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a","Type":"ContainerDied","Data":"2047006fc445345d89f058d9944e8a8d61a4d649ec8743ceee1c32242849733f"} Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.072221 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-jvrwg" Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.208964 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-config-data\") pod \"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a\" (UID: \"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a\") " Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.209085 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-job-config-data\") pod \"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a\" (UID: \"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a\") " Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.209147 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bg5lm\" (UniqueName: \"kubernetes.io/projected/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-kube-api-access-bg5lm\") pod \"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a\" (UID: \"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a\") " Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.209180 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-combined-ca-bundle\") pod \"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a\" (UID: \"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a\") " Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.215748 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-job-config-data" (OuterVolumeSpecName: "job-config-data") pod "1f88f6fb-7da5-47f6-b0b5-a33af1b6230a" (UID: "1f88f6fb-7da5-47f6-b0b5-a33af1b6230a"). InnerVolumeSpecName "job-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.218734 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-kube-api-access-bg5lm" (OuterVolumeSpecName: "kube-api-access-bg5lm") pod "1f88f6fb-7da5-47f6-b0b5-a33af1b6230a" (UID: "1f88f6fb-7da5-47f6-b0b5-a33af1b6230a"). InnerVolumeSpecName "kube-api-access-bg5lm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.219166 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-config-data" (OuterVolumeSpecName: "config-data") pod "1f88f6fb-7da5-47f6-b0b5-a33af1b6230a" (UID: "1f88f6fb-7da5-47f6-b0b5-a33af1b6230a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.242223 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1f88f6fb-7da5-47f6-b0b5-a33af1b6230a" (UID: "1f88f6fb-7da5-47f6-b0b5-a33af1b6230a"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.311994 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.312024 4894 reconciler_common.go:293] "Volume detached for volume \"job-config-data\" (UniqueName: \"kubernetes.io/secret/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-job-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.312036 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bg5lm\" (UniqueName: \"kubernetes.io/projected/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-kube-api-access-bg5lm\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.312044 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.497754 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-db-sync-jvrwg" event={"ID":"1f88f6fb-7da5-47f6-b0b5-a33af1b6230a","Type":"ContainerDied","Data":"3af203d1ff32e25047dcaf61fc68496d893d59fcca2dc3ca0ad571e813179d81"} Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.497803 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3af203d1ff32e25047dcaf61fc68496d893d59fcca2dc3ca0ad571e813179d81" Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.497860 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-db-sync-jvrwg" Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.959820 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Dec 09 16:22:35 crc kubenswrapper[4894]: E1209 16:22:35.960247 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f88f6fb-7da5-47f6-b0b5-a33af1b6230a" containerName="manila-db-sync" Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.960262 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f88f6fb-7da5-47f6-b0b5-a33af1b6230a" containerName="manila-db-sync" Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.960456 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f88f6fb-7da5-47f6-b0b5-a33af1b6230a" containerName="manila-db-sync" Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.961423 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.964042 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.964391 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scripts" Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.964447 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-manila-dockercfg-g2lbf" Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.964631 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-config-data" Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.988480 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.989980 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 09 16:22:35 crc kubenswrapper[4894]: I1209 16:22:35.992590 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.000151 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.024944 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.028402 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2b481c46-698a-4f16-88f1-1eef26dc1003-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.028441 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-scripts\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.028482 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-config-data\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.028505 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2b481c46-698a-4f16-88f1-1eef26dc1003-ceph\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.028541 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7nn2n\" (UniqueName: \"kubernetes.io/projected/2b481c46-698a-4f16-88f1-1eef26dc1003-kube-api-access-7nn2n\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.028574 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.028621 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2b481c46-698a-4f16-88f1-1eef26dc1003-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.028710 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.074887 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-67d475fdcf-c757l"] Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.076445 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67d475fdcf-c757l" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.094435 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67d475fdcf-c757l"] Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.130274 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28ef8474-844f-4525-8386-0e5cc06af8f7-dns-svc\") pod \"dnsmasq-dns-67d475fdcf-c757l\" (UID: \"28ef8474-844f-4525-8386-0e5cc06af8f7\") " pod="openstack/dnsmasq-dns-67d475fdcf-c757l" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.130325 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7nn2n\" (UniqueName: \"kubernetes.io/projected/2b481c46-698a-4f16-88f1-1eef26dc1003-kube-api-access-7nn2n\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.130351 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7jfzf\" (UniqueName: \"kubernetes.io/projected/28ef8474-844f-4525-8386-0e5cc06af8f7-kube-api-access-7jfzf\") pod \"dnsmasq-dns-67d475fdcf-c757l\" (UID: \"28ef8474-844f-4525-8386-0e5cc06af8f7\") " pod="openstack/dnsmasq-dns-67d475fdcf-c757l" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.130380 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-scripts\") pod \"manila-scheduler-0\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " pod="openstack/manila-scheduler-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.130400 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.130456 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2b481c46-698a-4f16-88f1-1eef26dc1003-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.130480 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28ef8474-844f-4525-8386-0e5cc06af8f7-config\") pod \"dnsmasq-dns-67d475fdcf-c757l\" (UID: \"28ef8474-844f-4525-8386-0e5cc06af8f7\") " pod="openstack/dnsmasq-dns-67d475fdcf-c757l" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.130501 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28ef8474-844f-4525-8386-0e5cc06af8f7-ovsdbserver-nb\") pod \"dnsmasq-dns-67d475fdcf-c757l\" (UID: \"28ef8474-844f-4525-8386-0e5cc06af8f7\") " pod="openstack/dnsmasq-dns-67d475fdcf-c757l" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.130520 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " pod="openstack/manila-scheduler-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.130548 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28ef8474-844f-4525-8386-0e5cc06af8f7-ovsdbserver-sb\") pod \"dnsmasq-dns-67d475fdcf-c757l\" (UID: \"28ef8474-844f-4525-8386-0e5cc06af8f7\") " pod="openstack/dnsmasq-dns-67d475fdcf-c757l" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.130567 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.130602 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2b481c46-698a-4f16-88f1-1eef26dc1003-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.130624 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-config-data\") pod \"manila-scheduler-0\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " pod="openstack/manila-scheduler-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.130655 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-scripts\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.130675 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " pod="openstack/manila-scheduler-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.130692 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbmwc\" (UniqueName: \"kubernetes.io/projected/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-kube-api-access-qbmwc\") pod \"manila-scheduler-0\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " pod="openstack/manila-scheduler-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.130708 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " pod="openstack/manila-scheduler-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.130724 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/28ef8474-844f-4525-8386-0e5cc06af8f7-openstack-edpm-ipam\") pod \"dnsmasq-dns-67d475fdcf-c757l\" (UID: \"28ef8474-844f-4525-8386-0e5cc06af8f7\") " pod="openstack/dnsmasq-dns-67d475fdcf-c757l" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.130754 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-config-data\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.130785 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2b481c46-698a-4f16-88f1-1eef26dc1003-ceph\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.131288 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2b481c46-698a-4f16-88f1-1eef26dc1003-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.132770 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2b481c46-698a-4f16-88f1-1eef26dc1003-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.136448 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.139146 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.139420 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-config-data\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.146179 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-scripts\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.150449 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2b481c46-698a-4f16-88f1-1eef26dc1003-ceph\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.153588 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7nn2n\" (UniqueName: \"kubernetes.io/projected/2b481c46-698a-4f16-88f1-1eef26dc1003-kube-api-access-7nn2n\") pod \"manila-share-share1-0\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.232331 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28ef8474-844f-4525-8386-0e5cc06af8f7-dns-svc\") pod \"dnsmasq-dns-67d475fdcf-c757l\" (UID: \"28ef8474-844f-4525-8386-0e5cc06af8f7\") " pod="openstack/dnsmasq-dns-67d475fdcf-c757l" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.232384 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7jfzf\" (UniqueName: \"kubernetes.io/projected/28ef8474-844f-4525-8386-0e5cc06af8f7-kube-api-access-7jfzf\") pod \"dnsmasq-dns-67d475fdcf-c757l\" (UID: \"28ef8474-844f-4525-8386-0e5cc06af8f7\") " pod="openstack/dnsmasq-dns-67d475fdcf-c757l" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.232414 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-scripts\") pod \"manila-scheduler-0\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " pod="openstack/manila-scheduler-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.232488 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28ef8474-844f-4525-8386-0e5cc06af8f7-config\") pod \"dnsmasq-dns-67d475fdcf-c757l\" (UID: \"28ef8474-844f-4525-8386-0e5cc06af8f7\") " pod="openstack/dnsmasq-dns-67d475fdcf-c757l" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.232514 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28ef8474-844f-4525-8386-0e5cc06af8f7-ovsdbserver-nb\") pod \"dnsmasq-dns-67d475fdcf-c757l\" (UID: \"28ef8474-844f-4525-8386-0e5cc06af8f7\") " pod="openstack/dnsmasq-dns-67d475fdcf-c757l" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.233187 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " pod="openstack/manila-scheduler-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.233217 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28ef8474-844f-4525-8386-0e5cc06af8f7-ovsdbserver-sb\") pod \"dnsmasq-dns-67d475fdcf-c757l\" (UID: \"28ef8474-844f-4525-8386-0e5cc06af8f7\") " pod="openstack/dnsmasq-dns-67d475fdcf-c757l" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.233300 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-config-data\") pod \"manila-scheduler-0\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " pod="openstack/manila-scheduler-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.233326 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " pod="openstack/manila-scheduler-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.233347 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbmwc\" (UniqueName: \"kubernetes.io/projected/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-kube-api-access-qbmwc\") pod \"manila-scheduler-0\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " pod="openstack/manila-scheduler-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.233362 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " pod="openstack/manila-scheduler-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.233379 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/28ef8474-844f-4525-8386-0e5cc06af8f7-openstack-edpm-ipam\") pod \"dnsmasq-dns-67d475fdcf-c757l\" (UID: \"28ef8474-844f-4525-8386-0e5cc06af8f7\") " pod="openstack/dnsmasq-dns-67d475fdcf-c757l" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.233387 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/28ef8474-844f-4525-8386-0e5cc06af8f7-config\") pod \"dnsmasq-dns-67d475fdcf-c757l\" (UID: \"28ef8474-844f-4525-8386-0e5cc06af8f7\") " pod="openstack/dnsmasq-dns-67d475fdcf-c757l" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.233453 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/28ef8474-844f-4525-8386-0e5cc06af8f7-ovsdbserver-nb\") pod \"dnsmasq-dns-67d475fdcf-c757l\" (UID: \"28ef8474-844f-4525-8386-0e5cc06af8f7\") " pod="openstack/dnsmasq-dns-67d475fdcf-c757l" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.233493 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/28ef8474-844f-4525-8386-0e5cc06af8f7-dns-svc\") pod \"dnsmasq-dns-67d475fdcf-c757l\" (UID: \"28ef8474-844f-4525-8386-0e5cc06af8f7\") " pod="openstack/dnsmasq-dns-67d475fdcf-c757l" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.233869 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " pod="openstack/manila-scheduler-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.233971 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/28ef8474-844f-4525-8386-0e5cc06af8f7-ovsdbserver-sb\") pod \"dnsmasq-dns-67d475fdcf-c757l\" (UID: \"28ef8474-844f-4525-8386-0e5cc06af8f7\") " pod="openstack/dnsmasq-dns-67d475fdcf-c757l" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.234242 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.234268 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/28ef8474-844f-4525-8386-0e5cc06af8f7-openstack-edpm-ipam\") pod \"dnsmasq-dns-67d475fdcf-c757l\" (UID: \"28ef8474-844f-4525-8386-0e5cc06af8f7\") " pod="openstack/dnsmasq-dns-67d475fdcf-c757l" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.235698 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.239833 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.240280 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-scripts\") pod \"manila-scheduler-0\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " pod="openstack/manila-scheduler-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.246930 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " pod="openstack/manila-scheduler-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.246955 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " pod="openstack/manila-scheduler-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.247598 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-config-data\") pod \"manila-scheduler-0\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " pod="openstack/manila-scheduler-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.251369 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbmwc\" (UniqueName: \"kubernetes.io/projected/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-kube-api-access-qbmwc\") pod \"manila-scheduler-0\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " pod="openstack/manila-scheduler-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.257250 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7jfzf\" (UniqueName: \"kubernetes.io/projected/28ef8474-844f-4525-8386-0e5cc06af8f7-kube-api-access-7jfzf\") pod \"dnsmasq-dns-67d475fdcf-c757l\" (UID: \"28ef8474-844f-4525-8386-0e5cc06af8f7\") " pod="openstack/dnsmasq-dns-67d475fdcf-c757l" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.260702 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.283310 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.333847 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.336468 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-config-data-custom\") pod \"manila-api-0\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.336521 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.336584 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-logs\") pod \"manila-api-0\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.336674 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-etc-machine-id\") pod \"manila-api-0\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.336703 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kwxtw\" (UniqueName: \"kubernetes.io/projected/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-kube-api-access-kwxtw\") pod \"manila-api-0\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.336723 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-scripts\") pod \"manila-api-0\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.336762 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-config-data\") pod \"manila-api-0\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.406063 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-67d475fdcf-c757l" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.442266 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-logs\") pod \"manila-api-0\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.442399 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-etc-machine-id\") pod \"manila-api-0\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.442430 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kwxtw\" (UniqueName: \"kubernetes.io/projected/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-kube-api-access-kwxtw\") pod \"manila-api-0\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.442467 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-scripts\") pod \"manila-api-0\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.442523 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-config-data\") pod \"manila-api-0\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.442564 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-config-data-custom\") pod \"manila-api-0\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.442590 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.444361 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-logs\") pod \"manila-api-0\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.444431 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-etc-machine-id\") pod \"manila-api-0\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.449751 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-config-data\") pod \"manila-api-0\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.451155 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-scripts\") pod \"manila-api-0\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.451884 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-config-data-custom\") pod \"manila-api-0\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.467761 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kwxtw\" (UniqueName: \"kubernetes.io/projected/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-kube-api-access-kwxtw\") pod \"manila-api-0\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.469407 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.685794 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.828588 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 09 16:22:36 crc kubenswrapper[4894]: W1209 16:22:36.841263 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8bb9ed46_f3f6_44fe_8fb2_359d97fc9c20.slice/crio-b7b6e447222e5341ffff4b7d251b3d3348677340047a11cfc1955f9ddfa93075 WatchSource:0}: Error finding container b7b6e447222e5341ffff4b7d251b3d3348677340047a11cfc1955f9ddfa93075: Status 404 returned error can't find the container with id b7b6e447222e5341ffff4b7d251b3d3348677340047a11cfc1955f9ddfa93075 Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.876283 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-67d475fdcf-c757l"] Dec 09 16:22:36 crc kubenswrapper[4894]: W1209 16:22:36.906818 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28ef8474_844f_4525_8386_0e5cc06af8f7.slice/crio-0646d2ae4549bb5b111f0b2763aaf342d148127f911cd98f984f0292574a19e4 WatchSource:0}: Error finding container 0646d2ae4549bb5b111f0b2763aaf342d148127f911cd98f984f0292574a19e4: Status 404 returned error can't find the container with id 0646d2ae4549bb5b111f0b2763aaf342d148127f911cd98f984f0292574a19e4 Dec 09 16:22:36 crc kubenswrapper[4894]: I1209 16:22:36.910109 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 09 16:22:36 crc kubenswrapper[4894]: W1209 16:22:36.938958 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2b481c46_698a_4f16_88f1_1eef26dc1003.slice/crio-bdb6380247ea1ec1b1a0abe9f6d4982c53c206fd466d786ec5cc9810c1985ca9 WatchSource:0}: Error finding container bdb6380247ea1ec1b1a0abe9f6d4982c53c206fd466d786ec5cc9810c1985ca9: Status 404 returned error can't find the container with id bdb6380247ea1ec1b1a0abe9f6d4982c53c206fd466d786ec5cc9810c1985ca9 Dec 09 16:22:37 crc kubenswrapper[4894]: I1209 16:22:37.307053 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 09 16:22:37 crc kubenswrapper[4894]: W1209 16:22:37.312623 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c6390f8_ba0c_49f9_a1af_b21cd0382f10.slice/crio-eab75a79db978e55dc9d7bbb70206fec0aa9f75201794e3c9ec93982afc34ac9 WatchSource:0}: Error finding container eab75a79db978e55dc9d7bbb70206fec0aa9f75201794e3c9ec93982afc34ac9: Status 404 returned error can't find the container with id eab75a79db978e55dc9d7bbb70206fec0aa9f75201794e3c9ec93982afc34ac9 Dec 09 16:22:37 crc kubenswrapper[4894]: I1209 16:22:37.545791 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"2b481c46-698a-4f16-88f1-1eef26dc1003","Type":"ContainerStarted","Data":"bdb6380247ea1ec1b1a0abe9f6d4982c53c206fd466d786ec5cc9810c1985ca9"} Dec 09 16:22:37 crc kubenswrapper[4894]: I1209 16:22:37.549477 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"5c6390f8-ba0c-49f9-a1af-b21cd0382f10","Type":"ContainerStarted","Data":"eab75a79db978e55dc9d7bbb70206fec0aa9f75201794e3c9ec93982afc34ac9"} Dec 09 16:22:37 crc kubenswrapper[4894]: I1209 16:22:37.550714 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20","Type":"ContainerStarted","Data":"b7b6e447222e5341ffff4b7d251b3d3348677340047a11cfc1955f9ddfa93075"} Dec 09 16:22:37 crc kubenswrapper[4894]: I1209 16:22:37.553744 4894 generic.go:334] "Generic (PLEG): container finished" podID="28ef8474-844f-4525-8386-0e5cc06af8f7" containerID="310e82d3a94b1edfad3faab1ef92aa438f64a408859df2d2a6da04d4a6658bb1" exitCode=0 Dec 09 16:22:37 crc kubenswrapper[4894]: I1209 16:22:37.553798 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67d475fdcf-c757l" event={"ID":"28ef8474-844f-4525-8386-0e5cc06af8f7","Type":"ContainerDied","Data":"310e82d3a94b1edfad3faab1ef92aa438f64a408859df2d2a6da04d4a6658bb1"} Dec 09 16:22:37 crc kubenswrapper[4894]: I1209 16:22:37.553827 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67d475fdcf-c757l" event={"ID":"28ef8474-844f-4525-8386-0e5cc06af8f7","Type":"ContainerStarted","Data":"0646d2ae4549bb5b111f0b2763aaf342d148127f911cd98f984f0292574a19e4"} Dec 09 16:22:38 crc kubenswrapper[4894]: I1209 16:22:38.571306 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20","Type":"ContainerStarted","Data":"ab2bf02aba8705478d62acb72b75d5d783a9b8d9f7aa0697510b8a93efbc6479"} Dec 09 16:22:38 crc kubenswrapper[4894]: I1209 16:22:38.580137 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-67d475fdcf-c757l" event={"ID":"28ef8474-844f-4525-8386-0e5cc06af8f7","Type":"ContainerStarted","Data":"ff98c9425ed61665097e983f475417ebc66cf9cde717875420cb27c874b317ad"} Dec 09 16:22:38 crc kubenswrapper[4894]: I1209 16:22:38.580283 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-67d475fdcf-c757l" Dec 09 16:22:38 crc kubenswrapper[4894]: I1209 16:22:38.582885 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"5c6390f8-ba0c-49f9-a1af-b21cd0382f10","Type":"ContainerStarted","Data":"4d4b7a70db7d8f72de6a2a957233a5ee4d375b0ff39ec49841c37338812ad425"} Dec 09 16:22:38 crc kubenswrapper[4894]: I1209 16:22:38.582919 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"5c6390f8-ba0c-49f9-a1af-b21cd0382f10","Type":"ContainerStarted","Data":"c232dbc52aa37fe5aa8edda4d34c602144ffd9e6ca79495d5449ada04965d899"} Dec 09 16:22:38 crc kubenswrapper[4894]: I1209 16:22:38.582983 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Dec 09 16:22:38 crc kubenswrapper[4894]: I1209 16:22:38.614214 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-67d475fdcf-c757l" podStartSLOduration=2.614188548 podStartE2EDuration="2.614188548s" podCreationTimestamp="2025-12-09 16:22:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:22:38.599385608 +0000 UTC m=+3052.918596277" watchObservedRunningTime="2025-12-09 16:22:38.614188548 +0000 UTC m=+3052.933399217" Dec 09 16:22:38 crc kubenswrapper[4894]: I1209 16:22:38.634010 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=2.633989153 podStartE2EDuration="2.633989153s" podCreationTimestamp="2025-12-09 16:22:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:22:38.625095583 +0000 UTC m=+3052.944306262" watchObservedRunningTime="2025-12-09 16:22:38.633989153 +0000 UTC m=+3052.953199822" Dec 09 16:22:38 crc kubenswrapper[4894]: I1209 16:22:38.954912 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Dec 09 16:22:39 crc kubenswrapper[4894]: I1209 16:22:39.602425 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20","Type":"ContainerStarted","Data":"da03e0cfc9740c6f6c9e7721da0c5c68f4fec02bab1f33e25dc9d40ef64acfdd"} Dec 09 16:22:39 crc kubenswrapper[4894]: I1209 16:22:39.629407 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.612567245 podStartE2EDuration="4.629383901s" podCreationTimestamp="2025-12-09 16:22:35 +0000 UTC" firstStartedPulling="2025-12-09 16:22:36.845954104 +0000 UTC m=+3051.165164773" lastFinishedPulling="2025-12-09 16:22:37.86277076 +0000 UTC m=+3052.181981429" observedRunningTime="2025-12-09 16:22:39.620424759 +0000 UTC m=+3053.939635448" watchObservedRunningTime="2025-12-09 16:22:39.629383901 +0000 UTC m=+3053.948594560" Dec 09 16:22:40 crc kubenswrapper[4894]: I1209 16:22:40.106460 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:22:40 crc kubenswrapper[4894]: E1209 16:22:40.106749 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:22:40 crc kubenswrapper[4894]: I1209 16:22:40.608519 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="5c6390f8-ba0c-49f9-a1af-b21cd0382f10" containerName="manila-api-log" containerID="cri-o://c232dbc52aa37fe5aa8edda4d34c602144ffd9e6ca79495d5449ada04965d899" gracePeriod=30 Dec 09 16:22:40 crc kubenswrapper[4894]: I1209 16:22:40.608692 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-api-0" podUID="5c6390f8-ba0c-49f9-a1af-b21cd0382f10" containerName="manila-api" containerID="cri-o://4d4b7a70db7d8f72de6a2a957233a5ee4d375b0ff39ec49841c37338812ad425" gracePeriod=30 Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.010391 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.010779 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" containerName="ceilometer-central-agent" containerID="cri-o://c72272b99978f0df125d44820f44392dfc1dc62bb9afde6b8726332b756ca55e" gracePeriod=30 Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.010912 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" containerName="sg-core" containerID="cri-o://97f3046a73e8242e7c85763c02199b24c68736ef49368531c0931f58272d480e" gracePeriod=30 Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.010948 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" containerName="ceilometer-notification-agent" containerID="cri-o://4c5103ffb3f6284c1b3040424af759fb01f8cc7149c20c54a01d1c212d3091b8" gracePeriod=30 Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.010929 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" containerName="proxy-httpd" containerID="cri-o://e2832eefa685d75df52c333784a4e0e53123f9dc331b58ea1d921d9235ece700" gracePeriod=30 Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.310449 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.458191 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-combined-ca-bundle\") pod \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.458243 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-scripts\") pod \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.458271 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-config-data-custom\") pod \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.458376 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-logs\") pod \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.458422 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-config-data\") pod \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.458540 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-etc-machine-id\") pod \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.458617 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kwxtw\" (UniqueName: \"kubernetes.io/projected/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-kube-api-access-kwxtw\") pod \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\" (UID: \"5c6390f8-ba0c-49f9-a1af-b21cd0382f10\") " Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.458795 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "5c6390f8-ba0c-49f9-a1af-b21cd0382f10" (UID: "5c6390f8-ba0c-49f9-a1af-b21cd0382f10"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.459079 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-logs" (OuterVolumeSpecName: "logs") pod "5c6390f8-ba0c-49f9-a1af-b21cd0382f10" (UID: "5c6390f8-ba0c-49f9-a1af-b21cd0382f10"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.459420 4894 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-logs\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.459445 4894 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.464797 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-scripts" (OuterVolumeSpecName: "scripts") pod "5c6390f8-ba0c-49f9-a1af-b21cd0382f10" (UID: "5c6390f8-ba0c-49f9-a1af-b21cd0382f10"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.467095 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-kube-api-access-kwxtw" (OuterVolumeSpecName: "kube-api-access-kwxtw") pod "5c6390f8-ba0c-49f9-a1af-b21cd0382f10" (UID: "5c6390f8-ba0c-49f9-a1af-b21cd0382f10"). InnerVolumeSpecName "kube-api-access-kwxtw". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.468405 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5c6390f8-ba0c-49f9-a1af-b21cd0382f10" (UID: "5c6390f8-ba0c-49f9-a1af-b21cd0382f10"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.497862 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c6390f8-ba0c-49f9-a1af-b21cd0382f10" (UID: "5c6390f8-ba0c-49f9-a1af-b21cd0382f10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.531732 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-config-data" (OuterVolumeSpecName: "config-data") pod "5c6390f8-ba0c-49f9-a1af-b21cd0382f10" (UID: "5c6390f8-ba0c-49f9-a1af-b21cd0382f10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.563290 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.563327 4894 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.563339 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.563348 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.563357 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kwxtw\" (UniqueName: \"kubernetes.io/projected/5c6390f8-ba0c-49f9-a1af-b21cd0382f10-kube-api-access-kwxtw\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.620919 4894 generic.go:334] "Generic (PLEG): container finished" podID="71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" containerID="e2832eefa685d75df52c333784a4e0e53123f9dc331b58ea1d921d9235ece700" exitCode=0 Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.621262 4894 generic.go:334] "Generic (PLEG): container finished" podID="71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" containerID="97f3046a73e8242e7c85763c02199b24c68736ef49368531c0931f58272d480e" exitCode=2 Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.621274 4894 generic.go:334] "Generic (PLEG): container finished" podID="71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" containerID="c72272b99978f0df125d44820f44392dfc1dc62bb9afde6b8726332b756ca55e" exitCode=0 Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.621318 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10","Type":"ContainerDied","Data":"e2832eefa685d75df52c333784a4e0e53123f9dc331b58ea1d921d9235ece700"} Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.621352 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10","Type":"ContainerDied","Data":"97f3046a73e8242e7c85763c02199b24c68736ef49368531c0931f58272d480e"} Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.621367 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10","Type":"ContainerDied","Data":"c72272b99978f0df125d44820f44392dfc1dc62bb9afde6b8726332b756ca55e"} Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.623339 4894 generic.go:334] "Generic (PLEG): container finished" podID="5c6390f8-ba0c-49f9-a1af-b21cd0382f10" containerID="4d4b7a70db7d8f72de6a2a957233a5ee4d375b0ff39ec49841c37338812ad425" exitCode=0 Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.623367 4894 generic.go:334] "Generic (PLEG): container finished" podID="5c6390f8-ba0c-49f9-a1af-b21cd0382f10" containerID="c232dbc52aa37fe5aa8edda4d34c602144ffd9e6ca79495d5449ada04965d899" exitCode=143 Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.623385 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"5c6390f8-ba0c-49f9-a1af-b21cd0382f10","Type":"ContainerDied","Data":"4d4b7a70db7d8f72de6a2a957233a5ee4d375b0ff39ec49841c37338812ad425"} Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.623406 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"5c6390f8-ba0c-49f9-a1af-b21cd0382f10","Type":"ContainerDied","Data":"c232dbc52aa37fe5aa8edda4d34c602144ffd9e6ca79495d5449ada04965d899"} Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.623419 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"5c6390f8-ba0c-49f9-a1af-b21cd0382f10","Type":"ContainerDied","Data":"eab75a79db978e55dc9d7bbb70206fec0aa9f75201794e3c9ec93982afc34ac9"} Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.623436 4894 scope.go:117] "RemoveContainer" containerID="4d4b7a70db7d8f72de6a2a957233a5ee4d375b0ff39ec49841c37338812ad425" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.623561 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.666375 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-api-0"] Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.688722 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-api-0"] Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.713206 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-api-0"] Dec 09 16:22:41 crc kubenswrapper[4894]: E1209 16:22:41.713691 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c6390f8-ba0c-49f9-a1af-b21cd0382f10" containerName="manila-api" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.713709 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c6390f8-ba0c-49f9-a1af-b21cd0382f10" containerName="manila-api" Dec 09 16:22:41 crc kubenswrapper[4894]: E1209 16:22:41.713727 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c6390f8-ba0c-49f9-a1af-b21cd0382f10" containerName="manila-api-log" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.713734 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c6390f8-ba0c-49f9-a1af-b21cd0382f10" containerName="manila-api-log" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.713938 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c6390f8-ba0c-49f9-a1af-b21cd0382f10" containerName="manila-api-log" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.713959 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c6390f8-ba0c-49f9-a1af-b21cd0382f10" containerName="manila-api" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.715050 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.718825 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-internal-svc" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.719007 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-manila-public-svc" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.720329 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-api-config-data" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.777762 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.867138 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/289ae5f1-46ed-4165-ac04-4c3cac996270-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.867190 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/289ae5f1-46ed-4165-ac04-4c3cac996270-etc-machine-id\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.867238 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/289ae5f1-46ed-4165-ac04-4c3cac996270-internal-tls-certs\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.867268 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/289ae5f1-46ed-4165-ac04-4c3cac996270-scripts\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.867349 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/289ae5f1-46ed-4165-ac04-4c3cac996270-config-data\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.867389 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/289ae5f1-46ed-4165-ac04-4c3cac996270-logs\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.867420 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nprkq\" (UniqueName: \"kubernetes.io/projected/289ae5f1-46ed-4165-ac04-4c3cac996270-kube-api-access-nprkq\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.867459 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/289ae5f1-46ed-4165-ac04-4c3cac996270-public-tls-certs\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.867517 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/289ae5f1-46ed-4165-ac04-4c3cac996270-config-data-custom\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.968954 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/289ae5f1-46ed-4165-ac04-4c3cac996270-config-data-custom\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.969729 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/289ae5f1-46ed-4165-ac04-4c3cac996270-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.969760 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/289ae5f1-46ed-4165-ac04-4c3cac996270-etc-machine-id\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.969791 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/289ae5f1-46ed-4165-ac04-4c3cac996270-internal-tls-certs\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.969813 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/289ae5f1-46ed-4165-ac04-4c3cac996270-scripts\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.969889 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/289ae5f1-46ed-4165-ac04-4c3cac996270-config-data\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.969926 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/289ae5f1-46ed-4165-ac04-4c3cac996270-logs\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.969971 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nprkq\" (UniqueName: \"kubernetes.io/projected/289ae5f1-46ed-4165-ac04-4c3cac996270-kube-api-access-nprkq\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.970006 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/289ae5f1-46ed-4165-ac04-4c3cac996270-public-tls-certs\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.971156 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/289ae5f1-46ed-4165-ac04-4c3cac996270-etc-machine-id\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.971981 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/289ae5f1-46ed-4165-ac04-4c3cac996270-logs\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.975245 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/289ae5f1-46ed-4165-ac04-4c3cac996270-config-data-custom\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.975905 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/289ae5f1-46ed-4165-ac04-4c3cac996270-config-data\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.976134 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/289ae5f1-46ed-4165-ac04-4c3cac996270-public-tls-certs\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.977762 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/289ae5f1-46ed-4165-ac04-4c3cac996270-scripts\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.977809 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/289ae5f1-46ed-4165-ac04-4c3cac996270-internal-tls-certs\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.997317 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/289ae5f1-46ed-4165-ac04-4c3cac996270-combined-ca-bundle\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:41 crc kubenswrapper[4894]: I1209 16:22:41.997951 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nprkq\" (UniqueName: \"kubernetes.io/projected/289ae5f1-46ed-4165-ac04-4c3cac996270-kube-api-access-nprkq\") pod \"manila-api-0\" (UID: \"289ae5f1-46ed-4165-ac04-4c3cac996270\") " pod="openstack/manila-api-0" Dec 09 16:22:42 crc kubenswrapper[4894]: I1209 16:22:42.099411 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-api-0" Dec 09 16:22:42 crc kubenswrapper[4894]: I1209 16:22:42.118641 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c6390f8-ba0c-49f9-a1af-b21cd0382f10" path="/var/lib/kubelet/pods/5c6390f8-ba0c-49f9-a1af-b21cd0382f10/volumes" Dec 09 16:22:43 crc kubenswrapper[4894]: I1209 16:22:43.644841 4894 generic.go:334] "Generic (PLEG): container finished" podID="71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" containerID="4c5103ffb3f6284c1b3040424af759fb01f8cc7149c20c54a01d1c212d3091b8" exitCode=0 Dec 09 16:22:43 crc kubenswrapper[4894]: I1209 16:22:43.644906 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10","Type":"ContainerDied","Data":"4c5103ffb3f6284c1b3040424af759fb01f8cc7149c20c54a01d1c212d3091b8"} Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.204239 4894 scope.go:117] "RemoveContainer" containerID="c232dbc52aa37fe5aa8edda4d34c602144ffd9e6ca79495d5449ada04965d899" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.449855 4894 scope.go:117] "RemoveContainer" containerID="4d4b7a70db7d8f72de6a2a957233a5ee4d375b0ff39ec49841c37338812ad425" Dec 09 16:22:44 crc kubenswrapper[4894]: E1209 16:22:44.452905 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4d4b7a70db7d8f72de6a2a957233a5ee4d375b0ff39ec49841c37338812ad425\": container with ID starting with 4d4b7a70db7d8f72de6a2a957233a5ee4d375b0ff39ec49841c37338812ad425 not found: ID does not exist" containerID="4d4b7a70db7d8f72de6a2a957233a5ee4d375b0ff39ec49841c37338812ad425" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.452964 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d4b7a70db7d8f72de6a2a957233a5ee4d375b0ff39ec49841c37338812ad425"} err="failed to get container status \"4d4b7a70db7d8f72de6a2a957233a5ee4d375b0ff39ec49841c37338812ad425\": rpc error: code = NotFound desc = could not find container \"4d4b7a70db7d8f72de6a2a957233a5ee4d375b0ff39ec49841c37338812ad425\": container with ID starting with 4d4b7a70db7d8f72de6a2a957233a5ee4d375b0ff39ec49841c37338812ad425 not found: ID does not exist" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.452991 4894 scope.go:117] "RemoveContainer" containerID="c232dbc52aa37fe5aa8edda4d34c602144ffd9e6ca79495d5449ada04965d899" Dec 09 16:22:44 crc kubenswrapper[4894]: E1209 16:22:44.453328 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c232dbc52aa37fe5aa8edda4d34c602144ffd9e6ca79495d5449ada04965d899\": container with ID starting with c232dbc52aa37fe5aa8edda4d34c602144ffd9e6ca79495d5449ada04965d899 not found: ID does not exist" containerID="c232dbc52aa37fe5aa8edda4d34c602144ffd9e6ca79495d5449ada04965d899" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.453382 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c232dbc52aa37fe5aa8edda4d34c602144ffd9e6ca79495d5449ada04965d899"} err="failed to get container status \"c232dbc52aa37fe5aa8edda4d34c602144ffd9e6ca79495d5449ada04965d899\": rpc error: code = NotFound desc = could not find container \"c232dbc52aa37fe5aa8edda4d34c602144ffd9e6ca79495d5449ada04965d899\": container with ID starting with c232dbc52aa37fe5aa8edda4d34c602144ffd9e6ca79495d5449ada04965d899 not found: ID does not exist" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.453399 4894 scope.go:117] "RemoveContainer" containerID="4d4b7a70db7d8f72de6a2a957233a5ee4d375b0ff39ec49841c37338812ad425" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.453881 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4d4b7a70db7d8f72de6a2a957233a5ee4d375b0ff39ec49841c37338812ad425"} err="failed to get container status \"4d4b7a70db7d8f72de6a2a957233a5ee4d375b0ff39ec49841c37338812ad425\": rpc error: code = NotFound desc = could not find container \"4d4b7a70db7d8f72de6a2a957233a5ee4d375b0ff39ec49841c37338812ad425\": container with ID starting with 4d4b7a70db7d8f72de6a2a957233a5ee4d375b0ff39ec49841c37338812ad425 not found: ID does not exist" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.453931 4894 scope.go:117] "RemoveContainer" containerID="c232dbc52aa37fe5aa8edda4d34c602144ffd9e6ca79495d5449ada04965d899" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.454526 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c232dbc52aa37fe5aa8edda4d34c602144ffd9e6ca79495d5449ada04965d899"} err="failed to get container status \"c232dbc52aa37fe5aa8edda4d34c602144ffd9e6ca79495d5449ada04965d899\": rpc error: code = NotFound desc = could not find container \"c232dbc52aa37fe5aa8edda4d34c602144ffd9e6ca79495d5449ada04965d899\": container with ID starting with c232dbc52aa37fe5aa8edda4d34c602144ffd9e6ca79495d5449ada04965d899 not found: ID does not exist" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.498306 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.633377 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-log-httpd\") pod \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.633925 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-combined-ca-bundle\") pod \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.633953 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-config-data\") pod \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.633979 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-run-httpd\") pod \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.634028 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-ceilometer-tls-certs\") pod \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.634054 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zx8l6\" (UniqueName: \"kubernetes.io/projected/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-kube-api-access-zx8l6\") pod \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.634110 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-sg-core-conf-yaml\") pod \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.634255 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-scripts\") pod \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\" (UID: \"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10\") " Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.634907 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" (UID: "71a4d0b7-457f-4bd0-92a7-3638c4c5ec10"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.635389 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" (UID: "71a4d0b7-457f-4bd0-92a7-3638c4c5ec10"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.644639 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-kube-api-access-zx8l6" (OuterVolumeSpecName: "kube-api-access-zx8l6") pod "71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" (UID: "71a4d0b7-457f-4bd0-92a7-3638c4c5ec10"). InnerVolumeSpecName "kube-api-access-zx8l6". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.645187 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-scripts" (OuterVolumeSpecName: "scripts") pod "71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" (UID: "71a4d0b7-457f-4bd0-92a7-3638c4c5ec10"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.660204 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"71a4d0b7-457f-4bd0-92a7-3638c4c5ec10","Type":"ContainerDied","Data":"92663ce5b12d2f73b12b4b15ff83d6702f0905e84c2901180df0ea041a6cb589"} Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.660226 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.660262 4894 scope.go:117] "RemoveContainer" containerID="e2832eefa685d75df52c333784a4e0e53123f9dc331b58ea1d921d9235ece700" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.681530 4894 scope.go:117] "RemoveContainer" containerID="97f3046a73e8242e7c85763c02199b24c68736ef49368531c0931f58272d480e" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.690290 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" (UID: "71a4d0b7-457f-4bd0-92a7-3638c4c5ec10"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.706349 4894 scope.go:117] "RemoveContainer" containerID="4c5103ffb3f6284c1b3040424af759fb01f8cc7149c20c54a01d1c212d3091b8" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.725369 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" (UID: "71a4d0b7-457f-4bd0-92a7-3638c4c5ec10"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.725793 4894 scope.go:117] "RemoveContainer" containerID="c72272b99978f0df125d44820f44392dfc1dc62bb9afde6b8726332b756ca55e" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.733667 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" (UID: "71a4d0b7-457f-4bd0-92a7-3638c4c5ec10"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.736723 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.736753 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.736762 4894 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.736787 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zx8l6\" (UniqueName: \"kubernetes.io/projected/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-kube-api-access-zx8l6\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.736797 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.736805 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.736814 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.772992 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-config-data" (OuterVolumeSpecName: "config-data") pod "71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" (UID: "71a4d0b7-457f-4bd0-92a7-3638c4c5ec10"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.774912 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-api-0"] Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.839018 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:44 crc kubenswrapper[4894]: I1209 16:22:44.995858 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.005797 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.029265 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 16:22:45 crc kubenswrapper[4894]: E1209 16:22:45.029770 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" containerName="ceilometer-notification-agent" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.029795 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" containerName="ceilometer-notification-agent" Dec 09 16:22:45 crc kubenswrapper[4894]: E1209 16:22:45.029822 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" containerName="sg-core" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.029832 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" containerName="sg-core" Dec 09 16:22:45 crc kubenswrapper[4894]: E1209 16:22:45.029849 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" containerName="ceilometer-central-agent" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.029857 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" containerName="ceilometer-central-agent" Dec 09 16:22:45 crc kubenswrapper[4894]: E1209 16:22:45.029885 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" containerName="proxy-httpd" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.029895 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" containerName="proxy-httpd" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.030159 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" containerName="ceilometer-notification-agent" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.030184 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" containerName="sg-core" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.030203 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" containerName="ceilometer-central-agent" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.030218 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" containerName="proxy-httpd" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.032387 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.034669 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.034736 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.035153 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.045090 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.144970 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7j4pg\" (UniqueName: \"kubernetes.io/projected/07937c4f-eb4d-4487-8ed4-bf9b55aba541-kube-api-access-7j4pg\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.145039 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-scripts\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.145060 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.145176 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.145200 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.145229 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07937c4f-eb4d-4487-8ed4-bf9b55aba541-log-httpd\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.145284 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07937c4f-eb4d-4487-8ed4-bf9b55aba541-run-httpd\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.145310 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-config-data\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.246883 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-scripts\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.246932 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.247053 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.247088 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.247126 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07937c4f-eb4d-4487-8ed4-bf9b55aba541-log-httpd\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.247188 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07937c4f-eb4d-4487-8ed4-bf9b55aba541-run-httpd\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.247218 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-config-data\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.247298 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7j4pg\" (UniqueName: \"kubernetes.io/projected/07937c4f-eb4d-4487-8ed4-bf9b55aba541-kube-api-access-7j4pg\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.248207 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07937c4f-eb4d-4487-8ed4-bf9b55aba541-log-httpd\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.248500 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07937c4f-eb4d-4487-8ed4-bf9b55aba541-run-httpd\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.252111 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.252110 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.253139 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-config-data\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.253560 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.254295 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-scripts\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.268536 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7j4pg\" (UniqueName: \"kubernetes.io/projected/07937c4f-eb4d-4487-8ed4-bf9b55aba541-kube-api-access-7j4pg\") pod \"ceilometer-0\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.351089 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.684480 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"289ae5f1-46ed-4165-ac04-4c3cac996270","Type":"ContainerStarted","Data":"6617560ed97635e3ce3c0ca15cff9832d1df6b3f5159fc2e55aadc524239d8e5"} Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.684850 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"289ae5f1-46ed-4165-ac04-4c3cac996270","Type":"ContainerStarted","Data":"2b5c4ff4699b3e532172169d43bb913f9059fd270e0a0425b1bf9e27bc4edc85"} Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.688266 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"2b481c46-698a-4f16-88f1-1eef26dc1003","Type":"ContainerStarted","Data":"82e3aea71db3514c9d84302cd834c96b6edce7a07cfd55c03222e0a761bc7aa9"} Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.688296 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"2b481c46-698a-4f16-88f1-1eef26dc1003","Type":"ContainerStarted","Data":"2e4c023204a180c3d82fea437de58e6116664edb7069c5dcdd7f45c70cc94b8a"} Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.720941 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=3.39070199 podStartE2EDuration="10.720918837s" podCreationTimestamp="2025-12-09 16:22:35 +0000 UTC" firstStartedPulling="2025-12-09 16:22:36.941650348 +0000 UTC m=+3051.260861007" lastFinishedPulling="2025-12-09 16:22:44.271867185 +0000 UTC m=+3058.591077854" observedRunningTime="2025-12-09 16:22:45.713519706 +0000 UTC m=+3060.032730375" watchObservedRunningTime="2025-12-09 16:22:45.720918837 +0000 UTC m=+3060.040129506" Dec 09 16:22:45 crc kubenswrapper[4894]: I1209 16:22:45.853153 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 16:22:46 crc kubenswrapper[4894]: I1209 16:22:46.120969 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="71a4d0b7-457f-4bd0-92a7-3638c4c5ec10" path="/var/lib/kubelet/pods/71a4d0b7-457f-4bd0-92a7-3638c4c5ec10/volumes" Dec 09 16:22:46 crc kubenswrapper[4894]: I1209 16:22:46.283411 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Dec 09 16:22:46 crc kubenswrapper[4894]: I1209 16:22:46.335235 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Dec 09 16:22:46 crc kubenswrapper[4894]: I1209 16:22:46.408808 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-67d475fdcf-c757l" Dec 09 16:22:46 crc kubenswrapper[4894]: I1209 16:22:46.609976 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dfc84b95f-nsxl6"] Dec 09 16:22:46 crc kubenswrapper[4894]: I1209 16:22:46.610517 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" podUID="ba44767a-6e2a-422d-84c1-17b60bb6620b" containerName="dnsmasq-dns" containerID="cri-o://55c2323af6293d64ad3288f86721689faf620a46518012bedee4f1828e01756f" gracePeriod=10 Dec 09 16:22:46 crc kubenswrapper[4894]: I1209 16:22:46.703131 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-api-0" event={"ID":"289ae5f1-46ed-4165-ac04-4c3cac996270","Type":"ContainerStarted","Data":"4932ce7de96d39454696d96b209fdcfa0a945f4c5c092748044f96365e17a427"} Dec 09 16:22:46 crc kubenswrapper[4894]: I1209 16:22:46.703197 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/manila-api-0" Dec 09 16:22:46 crc kubenswrapper[4894]: I1209 16:22:46.707118 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"07937c4f-eb4d-4487-8ed4-bf9b55aba541","Type":"ContainerStarted","Data":"7eec431f34bd10c4dc6282d79202f35c40c370ce4b18ea77808800b44caa8b9c"} Dec 09 16:22:46 crc kubenswrapper[4894]: I1209 16:22:46.727111 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-api-0" podStartSLOduration=5.727091895 podStartE2EDuration="5.727091895s" podCreationTimestamp="2025-12-09 16:22:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:22:46.71909708 +0000 UTC m=+3061.038307749" watchObservedRunningTime="2025-12-09 16:22:46.727091895 +0000 UTC m=+3061.046302564" Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.292437 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.422820 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-dns-svc\") pod \"ba44767a-6e2a-422d-84c1-17b60bb6620b\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.422912 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rpjpm\" (UniqueName: \"kubernetes.io/projected/ba44767a-6e2a-422d-84c1-17b60bb6620b-kube-api-access-rpjpm\") pod \"ba44767a-6e2a-422d-84c1-17b60bb6620b\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.422977 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-ovsdbserver-sb\") pod \"ba44767a-6e2a-422d-84c1-17b60bb6620b\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.422992 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-config\") pod \"ba44767a-6e2a-422d-84c1-17b60bb6620b\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.423552 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-openstack-edpm-ipam\") pod \"ba44767a-6e2a-422d-84c1-17b60bb6620b\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.423726 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-ovsdbserver-nb\") pod \"ba44767a-6e2a-422d-84c1-17b60bb6620b\" (UID: \"ba44767a-6e2a-422d-84c1-17b60bb6620b\") " Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.433850 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ba44767a-6e2a-422d-84c1-17b60bb6620b-kube-api-access-rpjpm" (OuterVolumeSpecName: "kube-api-access-rpjpm") pod "ba44767a-6e2a-422d-84c1-17b60bb6620b" (UID: "ba44767a-6e2a-422d-84c1-17b60bb6620b"). InnerVolumeSpecName "kube-api-access-rpjpm". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.476588 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-config" (OuterVolumeSpecName: "config") pod "ba44767a-6e2a-422d-84c1-17b60bb6620b" (UID: "ba44767a-6e2a-422d-84c1-17b60bb6620b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.482429 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-openstack-edpm-ipam" (OuterVolumeSpecName: "openstack-edpm-ipam") pod "ba44767a-6e2a-422d-84c1-17b60bb6620b" (UID: "ba44767a-6e2a-422d-84c1-17b60bb6620b"). InnerVolumeSpecName "openstack-edpm-ipam". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.484355 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "ba44767a-6e2a-422d-84c1-17b60bb6620b" (UID: "ba44767a-6e2a-422d-84c1-17b60bb6620b"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.488988 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "ba44767a-6e2a-422d-84c1-17b60bb6620b" (UID: "ba44767a-6e2a-422d-84c1-17b60bb6620b"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.495221 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "ba44767a-6e2a-422d-84c1-17b60bb6620b" (UID: "ba44767a-6e2a-422d-84c1-17b60bb6620b"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.526081 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-edpm-ipam\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-openstack-edpm-ipam\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.526325 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.526335 4894 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-dns-svc\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.526344 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rpjpm\" (UniqueName: \"kubernetes.io/projected/ba44767a-6e2a-422d-84c1-17b60bb6620b-kube-api-access-rpjpm\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.526353 4894 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.526363 4894 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ba44767a-6e2a-422d-84c1-17b60bb6620b-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.719774 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"07937c4f-eb4d-4487-8ed4-bf9b55aba541","Type":"ContainerStarted","Data":"228da1a824c1215b32b05aa5237aaa5ba91ae22d35f8d03711b8b6d460264daa"} Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.723104 4894 generic.go:334] "Generic (PLEG): container finished" podID="ba44767a-6e2a-422d-84c1-17b60bb6620b" containerID="55c2323af6293d64ad3288f86721689faf620a46518012bedee4f1828e01756f" exitCode=0 Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.723177 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.723168 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" event={"ID":"ba44767a-6e2a-422d-84c1-17b60bb6620b","Type":"ContainerDied","Data":"55c2323af6293d64ad3288f86721689faf620a46518012bedee4f1828e01756f"} Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.723244 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5dfc84b95f-nsxl6" event={"ID":"ba44767a-6e2a-422d-84c1-17b60bb6620b","Type":"ContainerDied","Data":"1204c89e5cb2d21c4e5aed06c10c2d632add2adb3cc16e446e7b4d8cf64b48b2"} Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.723270 4894 scope.go:117] "RemoveContainer" containerID="55c2323af6293d64ad3288f86721689faf620a46518012bedee4f1828e01756f" Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.779522 4894 scope.go:117] "RemoveContainer" containerID="55f1157d0f458b89f380f3c7ac519777160861ab1b12eff4c262aa4d8f204939" Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.783297 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5dfc84b95f-nsxl6"] Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.799691 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5dfc84b95f-nsxl6"] Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.817054 4894 scope.go:117] "RemoveContainer" containerID="55c2323af6293d64ad3288f86721689faf620a46518012bedee4f1828e01756f" Dec 09 16:22:47 crc kubenswrapper[4894]: E1209 16:22:47.817515 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55c2323af6293d64ad3288f86721689faf620a46518012bedee4f1828e01756f\": container with ID starting with 55c2323af6293d64ad3288f86721689faf620a46518012bedee4f1828e01756f not found: ID does not exist" containerID="55c2323af6293d64ad3288f86721689faf620a46518012bedee4f1828e01756f" Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.817552 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55c2323af6293d64ad3288f86721689faf620a46518012bedee4f1828e01756f"} err="failed to get container status \"55c2323af6293d64ad3288f86721689faf620a46518012bedee4f1828e01756f\": rpc error: code = NotFound desc = could not find container \"55c2323af6293d64ad3288f86721689faf620a46518012bedee4f1828e01756f\": container with ID starting with 55c2323af6293d64ad3288f86721689faf620a46518012bedee4f1828e01756f not found: ID does not exist" Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.817577 4894 scope.go:117] "RemoveContainer" containerID="55f1157d0f458b89f380f3c7ac519777160861ab1b12eff4c262aa4d8f204939" Dec 09 16:22:47 crc kubenswrapper[4894]: E1209 16:22:47.818571 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55f1157d0f458b89f380f3c7ac519777160861ab1b12eff4c262aa4d8f204939\": container with ID starting with 55f1157d0f458b89f380f3c7ac519777160861ab1b12eff4c262aa4d8f204939 not found: ID does not exist" containerID="55f1157d0f458b89f380f3c7ac519777160861ab1b12eff4c262aa4d8f204939" Dec 09 16:22:47 crc kubenswrapper[4894]: I1209 16:22:47.818665 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55f1157d0f458b89f380f3c7ac519777160861ab1b12eff4c262aa4d8f204939"} err="failed to get container status \"55f1157d0f458b89f380f3c7ac519777160861ab1b12eff4c262aa4d8f204939\": rpc error: code = NotFound desc = could not find container \"55f1157d0f458b89f380f3c7ac519777160861ab1b12eff4c262aa4d8f204939\": container with ID starting with 55f1157d0f458b89f380f3c7ac519777160861ab1b12eff4c262aa4d8f204939 not found: ID does not exist" Dec 09 16:22:48 crc kubenswrapper[4894]: I1209 16:22:48.119541 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ba44767a-6e2a-422d-84c1-17b60bb6620b" path="/var/lib/kubelet/pods/ba44767a-6e2a-422d-84c1-17b60bb6620b/volumes" Dec 09 16:22:48 crc kubenswrapper[4894]: I1209 16:22:48.733474 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"07937c4f-eb4d-4487-8ed4-bf9b55aba541","Type":"ContainerStarted","Data":"6b99ea2b2e4a8ac9a686ab6ae309ef3b88140b90ae86019bca4ba38ad992f6eb"} Dec 09 16:22:48 crc kubenswrapper[4894]: I1209 16:22:48.733513 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"07937c4f-eb4d-4487-8ed4-bf9b55aba541","Type":"ContainerStarted","Data":"2999f4db583377a8063ba25b986928f6b5092f0186c8de910cea657c35695674"} Dec 09 16:22:50 crc kubenswrapper[4894]: I1209 16:22:50.119915 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 16:22:50 crc kubenswrapper[4894]: I1209 16:22:50.755154 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"07937c4f-eb4d-4487-8ed4-bf9b55aba541","Type":"ContainerStarted","Data":"7e118c7416a0c910851a8d1dea860c2c015a132818086b80cc884f76269e4fda"} Dec 09 16:22:50 crc kubenswrapper[4894]: I1209 16:22:50.755468 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="07937c4f-eb4d-4487-8ed4-bf9b55aba541" containerName="proxy-httpd" containerID="cri-o://7e118c7416a0c910851a8d1dea860c2c015a132818086b80cc884f76269e4fda" gracePeriod=30 Dec 09 16:22:50 crc kubenswrapper[4894]: I1209 16:22:50.755412 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="07937c4f-eb4d-4487-8ed4-bf9b55aba541" containerName="ceilometer-central-agent" containerID="cri-o://228da1a824c1215b32b05aa5237aaa5ba91ae22d35f8d03711b8b6d460264daa" gracePeriod=30 Dec 09 16:22:50 crc kubenswrapper[4894]: I1209 16:22:50.755468 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="07937c4f-eb4d-4487-8ed4-bf9b55aba541" containerName="ceilometer-notification-agent" containerID="cri-o://2999f4db583377a8063ba25b986928f6b5092f0186c8de910cea657c35695674" gracePeriod=30 Dec 09 16:22:50 crc kubenswrapper[4894]: I1209 16:22:50.755511 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 16:22:50 crc kubenswrapper[4894]: I1209 16:22:50.755420 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="07937c4f-eb4d-4487-8ed4-bf9b55aba541" containerName="sg-core" containerID="cri-o://6b99ea2b2e4a8ac9a686ab6ae309ef3b88140b90ae86019bca4ba38ad992f6eb" gracePeriod=30 Dec 09 16:22:50 crc kubenswrapper[4894]: I1209 16:22:50.783270 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.262054674 podStartE2EDuration="6.783246041s" podCreationTimestamp="2025-12-09 16:22:44 +0000 UTC" firstStartedPulling="2025-12-09 16:22:45.865069391 +0000 UTC m=+3060.184280060" lastFinishedPulling="2025-12-09 16:22:50.386260748 +0000 UTC m=+3064.705471427" observedRunningTime="2025-12-09 16:22:50.772495362 +0000 UTC m=+3065.091706041" watchObservedRunningTime="2025-12-09 16:22:50.783246041 +0000 UTC m=+3065.102456720" Dec 09 16:22:51 crc kubenswrapper[4894]: I1209 16:22:51.767409 4894 generic.go:334] "Generic (PLEG): container finished" podID="07937c4f-eb4d-4487-8ed4-bf9b55aba541" containerID="6b99ea2b2e4a8ac9a686ab6ae309ef3b88140b90ae86019bca4ba38ad992f6eb" exitCode=2 Dec 09 16:22:51 crc kubenswrapper[4894]: I1209 16:22:51.767802 4894 generic.go:334] "Generic (PLEG): container finished" podID="07937c4f-eb4d-4487-8ed4-bf9b55aba541" containerID="2999f4db583377a8063ba25b986928f6b5092f0186c8de910cea657c35695674" exitCode=0 Dec 09 16:22:51 crc kubenswrapper[4894]: I1209 16:22:51.767468 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"07937c4f-eb4d-4487-8ed4-bf9b55aba541","Type":"ContainerDied","Data":"6b99ea2b2e4a8ac9a686ab6ae309ef3b88140b90ae86019bca4ba38ad992f6eb"} Dec 09 16:22:51 crc kubenswrapper[4894]: I1209 16:22:51.767861 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"07937c4f-eb4d-4487-8ed4-bf9b55aba541","Type":"ContainerDied","Data":"2999f4db583377a8063ba25b986928f6b5092f0186c8de910cea657c35695674"} Dec 09 16:22:53 crc kubenswrapper[4894]: I1209 16:22:53.788094 4894 generic.go:334] "Generic (PLEG): container finished" podID="07937c4f-eb4d-4487-8ed4-bf9b55aba541" containerID="228da1a824c1215b32b05aa5237aaa5ba91ae22d35f8d03711b8b6d460264daa" exitCode=0 Dec 09 16:22:53 crc kubenswrapper[4894]: I1209 16:22:53.788217 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"07937c4f-eb4d-4487-8ed4-bf9b55aba541","Type":"ContainerDied","Data":"228da1a824c1215b32b05aa5237aaa5ba91ae22d35f8d03711b8b6d460264daa"} Dec 09 16:22:55 crc kubenswrapper[4894]: I1209 16:22:55.107532 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:22:55 crc kubenswrapper[4894]: E1209 16:22:55.109811 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:22:57 crc kubenswrapper[4894]: I1209 16:22:57.827818 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Dec 09 16:22:57 crc kubenswrapper[4894]: I1209 16:22:57.888399 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Dec 09 16:22:57 crc kubenswrapper[4894]: I1209 16:22:57.903030 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Dec 09 16:22:57 crc kubenswrapper[4894]: I1209 16:22:57.931921 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Dec 09 16:22:58 crc kubenswrapper[4894]: I1209 16:22:58.840207 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20" containerName="manila-scheduler" containerID="cri-o://ab2bf02aba8705478d62acb72b75d5d783a9b8d9f7aa0697510b8a93efbc6479" gracePeriod=30 Dec 09 16:22:58 crc kubenswrapper[4894]: I1209 16:22:58.840248 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-scheduler-0" podUID="8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20" containerName="probe" containerID="cri-o://da03e0cfc9740c6f6c9e7721da0c5c68f4fec02bab1f33e25dc9d40ef64acfdd" gracePeriod=30 Dec 09 16:22:58 crc kubenswrapper[4894]: I1209 16:22:58.840376 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="2b481c46-698a-4f16-88f1-1eef26dc1003" containerName="manila-share" containerID="cri-o://2e4c023204a180c3d82fea437de58e6116664edb7069c5dcdd7f45c70cc94b8a" gracePeriod=30 Dec 09 16:22:58 crc kubenswrapper[4894]: I1209 16:22:58.840441 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/manila-share-share1-0" podUID="2b481c46-698a-4f16-88f1-1eef26dc1003" containerName="probe" containerID="cri-o://82e3aea71db3514c9d84302cd834c96b6edce7a07cfd55c03222e0a761bc7aa9" gracePeriod=30 Dec 09 16:22:59 crc kubenswrapper[4894]: I1209 16:22:59.856886 4894 generic.go:334] "Generic (PLEG): container finished" podID="8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20" containerID="da03e0cfc9740c6f6c9e7721da0c5c68f4fec02bab1f33e25dc9d40ef64acfdd" exitCode=0 Dec 09 16:22:59 crc kubenswrapper[4894]: I1209 16:22:59.856964 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20","Type":"ContainerDied","Data":"da03e0cfc9740c6f6c9e7721da0c5c68f4fec02bab1f33e25dc9d40ef64acfdd"} Dec 09 16:22:59 crc kubenswrapper[4894]: I1209 16:22:59.860469 4894 generic.go:334] "Generic (PLEG): container finished" podID="2b481c46-698a-4f16-88f1-1eef26dc1003" containerID="82e3aea71db3514c9d84302cd834c96b6edce7a07cfd55c03222e0a761bc7aa9" exitCode=0 Dec 09 16:22:59 crc kubenswrapper[4894]: I1209 16:22:59.860517 4894 generic.go:334] "Generic (PLEG): container finished" podID="2b481c46-698a-4f16-88f1-1eef26dc1003" containerID="2e4c023204a180c3d82fea437de58e6116664edb7069c5dcdd7f45c70cc94b8a" exitCode=1 Dec 09 16:22:59 crc kubenswrapper[4894]: I1209 16:22:59.860586 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"2b481c46-698a-4f16-88f1-1eef26dc1003","Type":"ContainerDied","Data":"82e3aea71db3514c9d84302cd834c96b6edce7a07cfd55c03222e0a761bc7aa9"} Dec 09 16:22:59 crc kubenswrapper[4894]: I1209 16:22:59.860633 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"2b481c46-698a-4f16-88f1-1eef26dc1003","Type":"ContainerDied","Data":"2e4c023204a180c3d82fea437de58e6116664edb7069c5dcdd7f45c70cc94b8a"} Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.006707 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.200804 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-scripts\") pod \"2b481c46-698a-4f16-88f1-1eef26dc1003\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.200899 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2b481c46-698a-4f16-88f1-1eef26dc1003-etc-machine-id\") pod \"2b481c46-698a-4f16-88f1-1eef26dc1003\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.200975 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2b481c46-698a-4f16-88f1-1eef26dc1003-ceph\") pod \"2b481c46-698a-4f16-88f1-1eef26dc1003\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.200998 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-config-data-custom\") pod \"2b481c46-698a-4f16-88f1-1eef26dc1003\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.201024 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2b481c46-698a-4f16-88f1-1eef26dc1003-var-lib-manila\") pod \"2b481c46-698a-4f16-88f1-1eef26dc1003\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.201054 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7nn2n\" (UniqueName: \"kubernetes.io/projected/2b481c46-698a-4f16-88f1-1eef26dc1003-kube-api-access-7nn2n\") pod \"2b481c46-698a-4f16-88f1-1eef26dc1003\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.201100 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-config-data\") pod \"2b481c46-698a-4f16-88f1-1eef26dc1003\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.201154 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b481c46-698a-4f16-88f1-1eef26dc1003-var-lib-manila" (OuterVolumeSpecName: "var-lib-manila") pod "2b481c46-698a-4f16-88f1-1eef26dc1003" (UID: "2b481c46-698a-4f16-88f1-1eef26dc1003"). InnerVolumeSpecName "var-lib-manila". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.201176 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-combined-ca-bundle\") pod \"2b481c46-698a-4f16-88f1-1eef26dc1003\" (UID: \"2b481c46-698a-4f16-88f1-1eef26dc1003\") " Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.202061 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2b481c46-698a-4f16-88f1-1eef26dc1003-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "2b481c46-698a-4f16-88f1-1eef26dc1003" (UID: "2b481c46-698a-4f16-88f1-1eef26dc1003"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.202560 4894 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/2b481c46-698a-4f16-88f1-1eef26dc1003-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.202751 4894 reconciler_common.go:293] "Volume detached for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/2b481c46-698a-4f16-88f1-1eef26dc1003-var-lib-manila\") on node \"crc\" DevicePath \"\"" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.207011 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "2b481c46-698a-4f16-88f1-1eef26dc1003" (UID: "2b481c46-698a-4f16-88f1-1eef26dc1003"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.208989 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b481c46-698a-4f16-88f1-1eef26dc1003-kube-api-access-7nn2n" (OuterVolumeSpecName: "kube-api-access-7nn2n") pod "2b481c46-698a-4f16-88f1-1eef26dc1003" (UID: "2b481c46-698a-4f16-88f1-1eef26dc1003"). InnerVolumeSpecName "kube-api-access-7nn2n". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.209473 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-scripts" (OuterVolumeSpecName: "scripts") pod "2b481c46-698a-4f16-88f1-1eef26dc1003" (UID: "2b481c46-698a-4f16-88f1-1eef26dc1003"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.218826 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2b481c46-698a-4f16-88f1-1eef26dc1003-ceph" (OuterVolumeSpecName: "ceph") pod "2b481c46-698a-4f16-88f1-1eef26dc1003" (UID: "2b481c46-698a-4f16-88f1-1eef26dc1003"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.257402 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2b481c46-698a-4f16-88f1-1eef26dc1003" (UID: "2b481c46-698a-4f16-88f1-1eef26dc1003"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.305103 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.305141 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.305153 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/2b481c46-698a-4f16-88f1-1eef26dc1003-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.305165 4894 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.305178 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7nn2n\" (UniqueName: \"kubernetes.io/projected/2b481c46-698a-4f16-88f1-1eef26dc1003-kube-api-access-7nn2n\") on node \"crc\" DevicePath \"\"" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.346876 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-config-data" (OuterVolumeSpecName: "config-data") pod "2b481c46-698a-4f16-88f1-1eef26dc1003" (UID: "2b481c46-698a-4f16-88f1-1eef26dc1003"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.406977 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2b481c46-698a-4f16-88f1-1eef26dc1003-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.720042 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.814341 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-etc-machine-id\") pod \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.814454 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20" (UID: "8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.814732 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbmwc\" (UniqueName: \"kubernetes.io/projected/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-kube-api-access-qbmwc\") pod \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.814848 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-combined-ca-bundle\") pod \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.814876 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-config-data\") pod \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.814955 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-config-data-custom\") pod \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.818905 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20" (UID: "8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.818952 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-scripts\") pod \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\" (UID: \"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20\") " Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.818974 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-kube-api-access-qbmwc" (OuterVolumeSpecName: "kube-api-access-qbmwc") pod "8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20" (UID: "8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20"). InnerVolumeSpecName "kube-api-access-qbmwc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.819586 4894 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-etc-machine-id\") on node \"crc\" DevicePath \"\"" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.819600 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbmwc\" (UniqueName: \"kubernetes.io/projected/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-kube-api-access-qbmwc\") on node \"crc\" DevicePath \"\"" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.819612 4894 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-config-data-custom\") on node \"crc\" DevicePath \"\"" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.823255 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-scripts" (OuterVolumeSpecName: "scripts") pod "8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20" (UID: "8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.870929 4894 generic.go:334] "Generic (PLEG): container finished" podID="8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20" containerID="ab2bf02aba8705478d62acb72b75d5d783a9b8d9f7aa0697510b8a93efbc6479" exitCode=0 Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.870977 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20","Type":"ContainerDied","Data":"ab2bf02aba8705478d62acb72b75d5d783a9b8d9f7aa0697510b8a93efbc6479"} Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.871001 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20","Type":"ContainerDied","Data":"b7b6e447222e5341ffff4b7d251b3d3348677340047a11cfc1955f9ddfa93075"} Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.871019 4894 scope.go:117] "RemoveContainer" containerID="da03e0cfc9740c6f6c9e7721da0c5c68f4fec02bab1f33e25dc9d40ef64acfdd" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.871127 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.875805 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"2b481c46-698a-4f16-88f1-1eef26dc1003","Type":"ContainerDied","Data":"bdb6380247ea1ec1b1a0abe9f6d4982c53c206fd466d786ec5cc9810c1985ca9"} Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.875879 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.882479 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20" (UID: "8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.904363 4894 scope.go:117] "RemoveContainer" containerID="ab2bf02aba8705478d62acb72b75d5d783a9b8d9f7aa0697510b8a93efbc6479" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.917755 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-share-share1-0"] Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.921855 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.921911 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.937528 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-share-share1-0"] Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.949568 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-share-share1-0"] Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.949864 4894 scope.go:117] "RemoveContainer" containerID="da03e0cfc9740c6f6c9e7721da0c5c68f4fec02bab1f33e25dc9d40ef64acfdd" Dec 09 16:23:00 crc kubenswrapper[4894]: E1209 16:23:00.950313 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da03e0cfc9740c6f6c9e7721da0c5c68f4fec02bab1f33e25dc9d40ef64acfdd\": container with ID starting with da03e0cfc9740c6f6c9e7721da0c5c68f4fec02bab1f33e25dc9d40ef64acfdd not found: ID does not exist" containerID="da03e0cfc9740c6f6c9e7721da0c5c68f4fec02bab1f33e25dc9d40ef64acfdd" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.950360 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da03e0cfc9740c6f6c9e7721da0c5c68f4fec02bab1f33e25dc9d40ef64acfdd"} err="failed to get container status \"da03e0cfc9740c6f6c9e7721da0c5c68f4fec02bab1f33e25dc9d40ef64acfdd\": rpc error: code = NotFound desc = could not find container \"da03e0cfc9740c6f6c9e7721da0c5c68f4fec02bab1f33e25dc9d40ef64acfdd\": container with ID starting with da03e0cfc9740c6f6c9e7721da0c5c68f4fec02bab1f33e25dc9d40ef64acfdd not found: ID does not exist" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.950385 4894 scope.go:117] "RemoveContainer" containerID="ab2bf02aba8705478d62acb72b75d5d783a9b8d9f7aa0697510b8a93efbc6479" Dec 09 16:23:00 crc kubenswrapper[4894]: E1209 16:23:00.950751 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b481c46-698a-4f16-88f1-1eef26dc1003" containerName="manila-share" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.950771 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b481c46-698a-4f16-88f1-1eef26dc1003" containerName="manila-share" Dec 09 16:23:00 crc kubenswrapper[4894]: E1209 16:23:00.950784 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20" containerName="manila-scheduler" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.950795 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20" containerName="manila-scheduler" Dec 09 16:23:00 crc kubenswrapper[4894]: E1209 16:23:00.950811 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20" containerName="probe" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.950819 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20" containerName="probe" Dec 09 16:23:00 crc kubenswrapper[4894]: E1209 16:23:00.950836 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba44767a-6e2a-422d-84c1-17b60bb6620b" containerName="init" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.950844 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba44767a-6e2a-422d-84c1-17b60bb6620b" containerName="init" Dec 09 16:23:00 crc kubenswrapper[4894]: E1209 16:23:00.950875 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2b481c46-698a-4f16-88f1-1eef26dc1003" containerName="probe" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.950885 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2b481c46-698a-4f16-88f1-1eef26dc1003" containerName="probe" Dec 09 16:23:00 crc kubenswrapper[4894]: E1209 16:23:00.950904 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ba44767a-6e2a-422d-84c1-17b60bb6620b" containerName="dnsmasq-dns" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.950912 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="ba44767a-6e2a-422d-84c1-17b60bb6620b" containerName="dnsmasq-dns" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.951137 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="ba44767a-6e2a-422d-84c1-17b60bb6620b" containerName="dnsmasq-dns" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.951167 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20" containerName="manila-scheduler" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.951201 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b481c46-698a-4f16-88f1-1eef26dc1003" containerName="manila-share" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.951214 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="2b481c46-698a-4f16-88f1-1eef26dc1003" containerName="probe" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.951233 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20" containerName="probe" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.952543 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 09 16:23:00 crc kubenswrapper[4894]: E1209 16:23:00.952938 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ab2bf02aba8705478d62acb72b75d5d783a9b8d9f7aa0697510b8a93efbc6479\": container with ID starting with ab2bf02aba8705478d62acb72b75d5d783a9b8d9f7aa0697510b8a93efbc6479 not found: ID does not exist" containerID="ab2bf02aba8705478d62acb72b75d5d783a9b8d9f7aa0697510b8a93efbc6479" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.952979 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ab2bf02aba8705478d62acb72b75d5d783a9b8d9f7aa0697510b8a93efbc6479"} err="failed to get container status \"ab2bf02aba8705478d62acb72b75d5d783a9b8d9f7aa0697510b8a93efbc6479\": rpc error: code = NotFound desc = could not find container \"ab2bf02aba8705478d62acb72b75d5d783a9b8d9f7aa0697510b8a93efbc6479\": container with ID starting with ab2bf02aba8705478d62acb72b75d5d783a9b8d9f7aa0697510b8a93efbc6479 not found: ID does not exist" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.953004 4894 scope.go:117] "RemoveContainer" containerID="82e3aea71db3514c9d84302cd834c96b6edce7a07cfd55c03222e0a761bc7aa9" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.957971 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.991583 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-share-share1-config-data" Dec 09 16:23:00 crc kubenswrapper[4894]: I1209 16:23:00.992957 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-config-data" (OuterVolumeSpecName: "config-data") pod "8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20" (UID: "8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.024708 4894 scope.go:117] "RemoveContainer" containerID="2e4c023204a180c3d82fea437de58e6116664edb7069c5dcdd7f45c70cc94b8a" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.028478 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/45d83e58-95c2-430b-bd84-1f1a0e05d512-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.028629 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/45d83e58-95c2-430b-bd84-1f1a0e05d512-ceph\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.028682 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45d83e58-95c2-430b-bd84-1f1a0e05d512-config-data\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.028754 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/45d83e58-95c2-430b-bd84-1f1a0e05d512-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.028794 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45d83e58-95c2-430b-bd84-1f1a0e05d512-scripts\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.028828 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqbl5\" (UniqueName: \"kubernetes.io/projected/45d83e58-95c2-430b-bd84-1f1a0e05d512-kube-api-access-vqbl5\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.028857 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/45d83e58-95c2-430b-bd84-1f1a0e05d512-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.028989 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45d83e58-95c2-430b-bd84-1f1a0e05d512-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.029590 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.131332 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/45d83e58-95c2-430b-bd84-1f1a0e05d512-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.131448 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/45d83e58-95c2-430b-bd84-1f1a0e05d512-etc-machine-id\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.131499 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45d83e58-95c2-430b-bd84-1f1a0e05d512-scripts\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.131569 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqbl5\" (UniqueName: \"kubernetes.io/projected/45d83e58-95c2-430b-bd84-1f1a0e05d512-kube-api-access-vqbl5\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.131623 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/45d83e58-95c2-430b-bd84-1f1a0e05d512-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.131723 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45d83e58-95c2-430b-bd84-1f1a0e05d512-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.131903 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-manila\" (UniqueName: \"kubernetes.io/host-path/45d83e58-95c2-430b-bd84-1f1a0e05d512-var-lib-manila\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.131992 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/45d83e58-95c2-430b-bd84-1f1a0e05d512-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.132228 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/45d83e58-95c2-430b-bd84-1f1a0e05d512-ceph\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.132268 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45d83e58-95c2-430b-bd84-1f1a0e05d512-config-data\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.136206 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/45d83e58-95c2-430b-bd84-1f1a0e05d512-combined-ca-bundle\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.136254 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/projected/45d83e58-95c2-430b-bd84-1f1a0e05d512-ceph\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.136758 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/45d83e58-95c2-430b-bd84-1f1a0e05d512-config-data-custom\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.136839 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/45d83e58-95c2-430b-bd84-1f1a0e05d512-config-data\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.138024 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/45d83e58-95c2-430b-bd84-1f1a0e05d512-scripts\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.147169 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqbl5\" (UniqueName: \"kubernetes.io/projected/45d83e58-95c2-430b-bd84-1f1a0e05d512-kube-api-access-vqbl5\") pod \"manila-share-share1-0\" (UID: \"45d83e58-95c2-430b-bd84-1f1a0e05d512\") " pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.209703 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-scheduler-0"] Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.228141 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-scheduler-0"] Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.239237 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/manila-scheduler-0"] Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.241414 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.243210 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"manila-scheduler-config-data" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.250682 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.309206 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-share-share1-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.336035 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ba40ccc-eccc-4780-b189-f849cce60042-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"1ba40ccc-eccc-4780-b189-f849cce60042\") " pod="openstack/manila-scheduler-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.336156 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ba40ccc-eccc-4780-b189-f849cce60042-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"1ba40ccc-eccc-4780-b189-f849cce60042\") " pod="openstack/manila-scheduler-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.336305 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ba40ccc-eccc-4780-b189-f849cce60042-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"1ba40ccc-eccc-4780-b189-f849cce60042\") " pod="openstack/manila-scheduler-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.336395 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ba40ccc-eccc-4780-b189-f849cce60042-scripts\") pod \"manila-scheduler-0\" (UID: \"1ba40ccc-eccc-4780-b189-f849cce60042\") " pod="openstack/manila-scheduler-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.336450 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rm77x\" (UniqueName: \"kubernetes.io/projected/1ba40ccc-eccc-4780-b189-f849cce60042-kube-api-access-rm77x\") pod \"manila-scheduler-0\" (UID: \"1ba40ccc-eccc-4780-b189-f849cce60042\") " pod="openstack/manila-scheduler-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.336555 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ba40ccc-eccc-4780-b189-f849cce60042-config-data\") pod \"manila-scheduler-0\" (UID: \"1ba40ccc-eccc-4780-b189-f849cce60042\") " pod="openstack/manila-scheduler-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.438767 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ba40ccc-eccc-4780-b189-f849cce60042-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"1ba40ccc-eccc-4780-b189-f849cce60042\") " pod="openstack/manila-scheduler-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.439238 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ba40ccc-eccc-4780-b189-f849cce60042-scripts\") pod \"manila-scheduler-0\" (UID: \"1ba40ccc-eccc-4780-b189-f849cce60042\") " pod="openstack/manila-scheduler-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.439397 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rm77x\" (UniqueName: \"kubernetes.io/projected/1ba40ccc-eccc-4780-b189-f849cce60042-kube-api-access-rm77x\") pod \"manila-scheduler-0\" (UID: \"1ba40ccc-eccc-4780-b189-f849cce60042\") " pod="openstack/manila-scheduler-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.439579 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ba40ccc-eccc-4780-b189-f849cce60042-config-data\") pod \"manila-scheduler-0\" (UID: \"1ba40ccc-eccc-4780-b189-f849cce60042\") " pod="openstack/manila-scheduler-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.439838 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ba40ccc-eccc-4780-b189-f849cce60042-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"1ba40ccc-eccc-4780-b189-f849cce60042\") " pod="openstack/manila-scheduler-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.440081 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ba40ccc-eccc-4780-b189-f849cce60042-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"1ba40ccc-eccc-4780-b189-f849cce60042\") " pod="openstack/manila-scheduler-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.440629 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1ba40ccc-eccc-4780-b189-f849cce60042-etc-machine-id\") pod \"manila-scheduler-0\" (UID: \"1ba40ccc-eccc-4780-b189-f849cce60042\") " pod="openstack/manila-scheduler-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.445334 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/1ba40ccc-eccc-4780-b189-f849cce60042-config-data-custom\") pod \"manila-scheduler-0\" (UID: \"1ba40ccc-eccc-4780-b189-f849cce60042\") " pod="openstack/manila-scheduler-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.450418 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1ba40ccc-eccc-4780-b189-f849cce60042-combined-ca-bundle\") pod \"manila-scheduler-0\" (UID: \"1ba40ccc-eccc-4780-b189-f849cce60042\") " pod="openstack/manila-scheduler-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.454402 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ba40ccc-eccc-4780-b189-f849cce60042-config-data\") pod \"manila-scheduler-0\" (UID: \"1ba40ccc-eccc-4780-b189-f849cce60042\") " pod="openstack/manila-scheduler-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.461366 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rm77x\" (UniqueName: \"kubernetes.io/projected/1ba40ccc-eccc-4780-b189-f849cce60042-kube-api-access-rm77x\") pod \"manila-scheduler-0\" (UID: \"1ba40ccc-eccc-4780-b189-f849cce60042\") " pod="openstack/manila-scheduler-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.463050 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ba40ccc-eccc-4780-b189-f849cce60042-scripts\") pod \"manila-scheduler-0\" (UID: \"1ba40ccc-eccc-4780-b189-f849cce60042\") " pod="openstack/manila-scheduler-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.624556 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/manila-scheduler-0" Dec 09 16:23:01 crc kubenswrapper[4894]: I1209 16:23:01.884452 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-share-share1-0"] Dec 09 16:23:01 crc kubenswrapper[4894]: W1209 16:23:01.890114 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod45d83e58_95c2_430b_bd84_1f1a0e05d512.slice/crio-6c012a81f4a9aa22b6de9cb3555cbcd517112ef083b752fee3af775fadf1d2b2 WatchSource:0}: Error finding container 6c012a81f4a9aa22b6de9cb3555cbcd517112ef083b752fee3af775fadf1d2b2: Status 404 returned error can't find the container with id 6c012a81f4a9aa22b6de9cb3555cbcd517112ef083b752fee3af775fadf1d2b2 Dec 09 16:23:02 crc kubenswrapper[4894]: I1209 16:23:02.131979 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2b481c46-698a-4f16-88f1-1eef26dc1003" path="/var/lib/kubelet/pods/2b481c46-698a-4f16-88f1-1eef26dc1003/volumes" Dec 09 16:23:02 crc kubenswrapper[4894]: I1209 16:23:02.133197 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20" path="/var/lib/kubelet/pods/8bb9ed46-f3f6-44fe-8fb2-359d97fc9c20/volumes" Dec 09 16:23:02 crc kubenswrapper[4894]: I1209 16:23:02.159369 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/manila-scheduler-0"] Dec 09 16:23:02 crc kubenswrapper[4894]: I1209 16:23:02.921954 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"45d83e58-95c2-430b-bd84-1f1a0e05d512","Type":"ContainerStarted","Data":"2cf380c34b8a5cb207151dbe38e068359b7dc691db191f8b39c92f07f471f202"} Dec 09 16:23:02 crc kubenswrapper[4894]: I1209 16:23:02.922492 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"45d83e58-95c2-430b-bd84-1f1a0e05d512","Type":"ContainerStarted","Data":"6c012a81f4a9aa22b6de9cb3555cbcd517112ef083b752fee3af775fadf1d2b2"} Dec 09 16:23:02 crc kubenswrapper[4894]: I1209 16:23:02.960857 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"1ba40ccc-eccc-4780-b189-f849cce60042","Type":"ContainerStarted","Data":"6918e7cf04ab0ee746b9d123560c5fa6906d42b9d0a37fe38b79fc9c6a710ca7"} Dec 09 16:23:02 crc kubenswrapper[4894]: I1209 16:23:02.960916 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"1ba40ccc-eccc-4780-b189-f849cce60042","Type":"ContainerStarted","Data":"674fef4b6bd796dac96e9f8e60310a8e50419fdca11a7f3272c5e7ecc738c634"} Dec 09 16:23:03 crc kubenswrapper[4894]: I1209 16:23:03.648815 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/manila-api-0" Dec 09 16:23:03 crc kubenswrapper[4894]: I1209 16:23:03.976606 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-share-share1-0" event={"ID":"45d83e58-95c2-430b-bd84-1f1a0e05d512","Type":"ContainerStarted","Data":"271e0949a14b31afb25515cff9206ab94113831eccbd78afd3c03e39e7db72ea"} Dec 09 16:23:03 crc kubenswrapper[4894]: I1209 16:23:03.978562 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/manila-scheduler-0" event={"ID":"1ba40ccc-eccc-4780-b189-f849cce60042","Type":"ContainerStarted","Data":"0f5585375b195ce040dc0390a8c5a05eb604ddc55ec824c275112361986b8f4e"} Dec 09 16:23:04 crc kubenswrapper[4894]: I1209 16:23:04.004908 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-share-share1-0" podStartSLOduration=4.004889457 podStartE2EDuration="4.004889457s" podCreationTimestamp="2025-12-09 16:23:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:23:03.998006111 +0000 UTC m=+3078.317216780" watchObservedRunningTime="2025-12-09 16:23:04.004889457 +0000 UTC m=+3078.324100126" Dec 09 16:23:04 crc kubenswrapper[4894]: I1209 16:23:04.026606 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/manila-scheduler-0" podStartSLOduration=3.026586844 podStartE2EDuration="3.026586844s" podCreationTimestamp="2025-12-09 16:23:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:23:04.017004805 +0000 UTC m=+3078.336215484" watchObservedRunningTime="2025-12-09 16:23:04.026586844 +0000 UTC m=+3078.345797513" Dec 09 16:23:09 crc kubenswrapper[4894]: I1209 16:23:09.108726 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:23:09 crc kubenswrapper[4894]: E1209 16:23:09.110097 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:23:11 crc kubenswrapper[4894]: I1209 16:23:11.310016 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-share-share1-0" Dec 09 16:23:11 crc kubenswrapper[4894]: I1209 16:23:11.625003 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/manila-scheduler-0" Dec 09 16:23:15 crc kubenswrapper[4894]: I1209 16:23:15.360305 4894 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ceilometer-0" podUID="07937c4f-eb4d-4487-8ed4-bf9b55aba541" containerName="proxy-httpd" probeResult="failure" output="HTTP probe failed with statuscode: 503" Dec 09 16:23:20 crc kubenswrapper[4894]: I1209 16:23:20.107733 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:23:20 crc kubenswrapper[4894]: E1209 16:23:20.108849 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.160798 4894 generic.go:334] "Generic (PLEG): container finished" podID="07937c4f-eb4d-4487-8ed4-bf9b55aba541" containerID="7e118c7416a0c910851a8d1dea860c2c015a132818086b80cc884f76269e4fda" exitCode=137 Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.161170 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"07937c4f-eb4d-4487-8ed4-bf9b55aba541","Type":"ContainerDied","Data":"7e118c7416a0c910851a8d1dea860c2c015a132818086b80cc884f76269e4fda"} Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.161201 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"07937c4f-eb4d-4487-8ed4-bf9b55aba541","Type":"ContainerDied","Data":"7eec431f34bd10c4dc6282d79202f35c40c370ce4b18ea77808800b44caa8b9c"} Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.161215 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7eec431f34bd10c4dc6282d79202f35c40c370ce4b18ea77808800b44caa8b9c" Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.193927 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.362452 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-sg-core-conf-yaml\") pod \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.362495 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-combined-ca-bundle\") pod \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.362577 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-config-data\") pod \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.363305 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07937c4f-eb4d-4487-8ed4-bf9b55aba541-log-httpd\") pod \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.363379 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-ceilometer-tls-certs\") pod \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.363401 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07937c4f-eb4d-4487-8ed4-bf9b55aba541-run-httpd\") pod \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.363495 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7j4pg\" (UniqueName: \"kubernetes.io/projected/07937c4f-eb4d-4487-8ed4-bf9b55aba541-kube-api-access-7j4pg\") pod \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.363544 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-scripts\") pod \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\" (UID: \"07937c4f-eb4d-4487-8ed4-bf9b55aba541\") " Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.364509 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07937c4f-eb4d-4487-8ed4-bf9b55aba541-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "07937c4f-eb4d-4487-8ed4-bf9b55aba541" (UID: "07937c4f-eb4d-4487-8ed4-bf9b55aba541"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.364984 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/07937c4f-eb4d-4487-8ed4-bf9b55aba541-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "07937c4f-eb4d-4487-8ed4-bf9b55aba541" (UID: "07937c4f-eb4d-4487-8ed4-bf9b55aba541"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.383016 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/07937c4f-eb4d-4487-8ed4-bf9b55aba541-kube-api-access-7j4pg" (OuterVolumeSpecName: "kube-api-access-7j4pg") pod "07937c4f-eb4d-4487-8ed4-bf9b55aba541" (UID: "07937c4f-eb4d-4487-8ed4-bf9b55aba541"). InnerVolumeSpecName "kube-api-access-7j4pg". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.386034 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-scripts" (OuterVolumeSpecName: "scripts") pod "07937c4f-eb4d-4487-8ed4-bf9b55aba541" (UID: "07937c4f-eb4d-4487-8ed4-bf9b55aba541"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.450466 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "07937c4f-eb4d-4487-8ed4-bf9b55aba541" (UID: "07937c4f-eb4d-4487-8ed4-bf9b55aba541"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.459225 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "07937c4f-eb4d-4487-8ed4-bf9b55aba541" (UID: "07937c4f-eb4d-4487-8ed4-bf9b55aba541"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.465932 4894 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07937c4f-eb4d-4487-8ed4-bf9b55aba541-log-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.465960 4894 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.465970 4894 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/07937c4f-eb4d-4487-8ed4-bf9b55aba541-run-httpd\") on node \"crc\" DevicePath \"\"" Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.465979 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7j4pg\" (UniqueName: \"kubernetes.io/projected/07937c4f-eb4d-4487-8ed4-bf9b55aba541-kube-api-access-7j4pg\") on node \"crc\" DevicePath \"\"" Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.465987 4894 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-scripts\") on node \"crc\" DevicePath \"\"" Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.465995 4894 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.467198 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "07937c4f-eb4d-4487-8ed4-bf9b55aba541" (UID: "07937c4f-eb4d-4487-8ed4-bf9b55aba541"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.505153 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-config-data" (OuterVolumeSpecName: "config-data") pod "07937c4f-eb4d-4487-8ed4-bf9b55aba541" (UID: "07937c4f-eb4d-4487-8ed4-bf9b55aba541"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.568705 4894 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Dec 09 16:23:21 crc kubenswrapper[4894]: I1209 16:23:21.568741 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/07937c4f-eb4d-4487-8ed4-bf9b55aba541-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.168673 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.195222 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.206897 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.240497 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Dec 09 16:23:22 crc kubenswrapper[4894]: E1209 16:23:22.241236 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07937c4f-eb4d-4487-8ed4-bf9b55aba541" containerName="proxy-httpd" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.241259 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="07937c4f-eb4d-4487-8ed4-bf9b55aba541" containerName="proxy-httpd" Dec 09 16:23:22 crc kubenswrapper[4894]: E1209 16:23:22.241297 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07937c4f-eb4d-4487-8ed4-bf9b55aba541" containerName="sg-core" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.241306 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="07937c4f-eb4d-4487-8ed4-bf9b55aba541" containerName="sg-core" Dec 09 16:23:22 crc kubenswrapper[4894]: E1209 16:23:22.241330 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07937c4f-eb4d-4487-8ed4-bf9b55aba541" containerName="ceilometer-notification-agent" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.241339 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="07937c4f-eb4d-4487-8ed4-bf9b55aba541" containerName="ceilometer-notification-agent" Dec 09 16:23:22 crc kubenswrapper[4894]: E1209 16:23:22.241702 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="07937c4f-eb4d-4487-8ed4-bf9b55aba541" containerName="ceilometer-central-agent" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.241720 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="07937c4f-eb4d-4487-8ed4-bf9b55aba541" containerName="ceilometer-central-agent" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.242248 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="07937c4f-eb4d-4487-8ed4-bf9b55aba541" containerName="sg-core" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.242289 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="07937c4f-eb4d-4487-8ed4-bf9b55aba541" containerName="ceilometer-notification-agent" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.242315 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="07937c4f-eb4d-4487-8ed4-bf9b55aba541" containerName="proxy-httpd" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.242331 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="07937c4f-eb4d-4487-8ed4-bf9b55aba541" containerName="ceilometer-central-agent" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.248185 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.252050 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.252559 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.253235 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.287324 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.388953 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.389020 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-scripts\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.389087 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-config-data\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.389165 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5w5c8\" (UniqueName: \"kubernetes.io/projected/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-kube-api-access-5w5c8\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.389254 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-log-httpd\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.389272 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-run-httpd\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.389338 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.389355 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.491376 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5w5c8\" (UniqueName: \"kubernetes.io/projected/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-kube-api-access-5w5c8\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.491480 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-log-httpd\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.491516 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-run-httpd\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.491657 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.491697 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.491786 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.491817 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-scripts\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.491865 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-config-data\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.492246 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-log-httpd\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.492250 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-run-httpd\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.496103 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.497412 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.498084 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-config-data\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.505255 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.509830 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-scripts\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.512607 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5w5c8\" (UniqueName: \"kubernetes.io/projected/c2ab328c-f7ef-432f-8e89-e7201fa7ad2c-kube-api-access-5w5c8\") pod \"ceilometer-0\" (UID: \"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c\") " pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.582343 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Dec 09 16:23:22 crc kubenswrapper[4894]: I1209 16:23:22.823273 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-share-share1-0" Dec 09 16:23:23 crc kubenswrapper[4894]: I1209 16:23:23.087376 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Dec 09 16:23:23 crc kubenswrapper[4894]: I1209 16:23:23.178906 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c","Type":"ContainerStarted","Data":"9c3cceb8293b4807d14bb52245c05659dbbe46cdbf8d76d1af121bde830c97a4"} Dec 09 16:23:23 crc kubenswrapper[4894]: I1209 16:23:23.400291 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/manila-scheduler-0" Dec 09 16:23:24 crc kubenswrapper[4894]: I1209 16:23:24.121530 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="07937c4f-eb4d-4487-8ed4-bf9b55aba541" path="/var/lib/kubelet/pods/07937c4f-eb4d-4487-8ed4-bf9b55aba541/volumes" Dec 09 16:23:24 crc kubenswrapper[4894]: I1209 16:23:24.203710 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c","Type":"ContainerStarted","Data":"4c2f6f8a2e751247a3cad685151800aef53bcb83f52eabc1337611edfc8113b2"} Dec 09 16:23:25 crc kubenswrapper[4894]: I1209 16:23:25.216480 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c","Type":"ContainerStarted","Data":"888cf5b296bd302b5a07ddd536d33f1b4705ff1484cb9273d6eec855a6d3cc42"} Dec 09 16:23:26 crc kubenswrapper[4894]: I1209 16:23:26.228217 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c","Type":"ContainerStarted","Data":"b6c53020ccfb8de50dcf32d1bdac8d0ce9eaf2a66274ac49f5cffcfef06875ba"} Dec 09 16:23:27 crc kubenswrapper[4894]: I1209 16:23:27.238100 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c2ab328c-f7ef-432f-8e89-e7201fa7ad2c","Type":"ContainerStarted","Data":"e1f767ebaddd835ce06deae8597da3db7cab306047de6bcb9e9dc9a14f98e42c"} Dec 09 16:23:27 crc kubenswrapper[4894]: I1209 16:23:27.238757 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Dec 09 16:23:27 crc kubenswrapper[4894]: I1209 16:23:27.269148 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.133127105 podStartE2EDuration="5.269130796s" podCreationTimestamp="2025-12-09 16:23:22 +0000 UTC" firstStartedPulling="2025-12-09 16:23:23.092350122 +0000 UTC m=+3097.411560791" lastFinishedPulling="2025-12-09 16:23:26.228353793 +0000 UTC m=+3100.547564482" observedRunningTime="2025-12-09 16:23:27.266268699 +0000 UTC m=+3101.585479368" watchObservedRunningTime="2025-12-09 16:23:27.269130796 +0000 UTC m=+3101.588341465" Dec 09 16:23:35 crc kubenswrapper[4894]: I1209 16:23:35.107891 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:23:35 crc kubenswrapper[4894]: E1209 16:23:35.108977 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:23:47 crc kubenswrapper[4894]: I1209 16:23:47.106699 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:23:47 crc kubenswrapper[4894]: E1209 16:23:47.107690 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:23:52 crc kubenswrapper[4894]: I1209 16:23:52.598044 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Dec 09 16:24:00 crc kubenswrapper[4894]: I1209 16:24:00.106313 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:24:00 crc kubenswrapper[4894]: E1209 16:24:00.107170 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:24:15 crc kubenswrapper[4894]: I1209 16:24:15.107535 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:24:15 crc kubenswrapper[4894]: E1209 16:24:15.108849 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:24:19 crc kubenswrapper[4894]: I1209 16:24:19.538755 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-lnlst"] Dec 09 16:24:19 crc kubenswrapper[4894]: I1209 16:24:19.544685 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lnlst" Dec 09 16:24:19 crc kubenswrapper[4894]: I1209 16:24:19.550235 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lnlst"] Dec 09 16:24:19 crc kubenswrapper[4894]: I1209 16:24:19.646224 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5sc2\" (UniqueName: \"kubernetes.io/projected/74ee2091-d89d-4945-af03-5b9887bdfa24-kube-api-access-x5sc2\") pod \"certified-operators-lnlst\" (UID: \"74ee2091-d89d-4945-af03-5b9887bdfa24\") " pod="openshift-marketplace/certified-operators-lnlst" Dec 09 16:24:19 crc kubenswrapper[4894]: I1209 16:24:19.647803 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74ee2091-d89d-4945-af03-5b9887bdfa24-catalog-content\") pod \"certified-operators-lnlst\" (UID: \"74ee2091-d89d-4945-af03-5b9887bdfa24\") " pod="openshift-marketplace/certified-operators-lnlst" Dec 09 16:24:19 crc kubenswrapper[4894]: I1209 16:24:19.647939 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74ee2091-d89d-4945-af03-5b9887bdfa24-utilities\") pod \"certified-operators-lnlst\" (UID: \"74ee2091-d89d-4945-af03-5b9887bdfa24\") " pod="openshift-marketplace/certified-operators-lnlst" Dec 09 16:24:19 crc kubenswrapper[4894]: I1209 16:24:19.750698 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74ee2091-d89d-4945-af03-5b9887bdfa24-utilities\") pod \"certified-operators-lnlst\" (UID: \"74ee2091-d89d-4945-af03-5b9887bdfa24\") " pod="openshift-marketplace/certified-operators-lnlst" Dec 09 16:24:19 crc kubenswrapper[4894]: I1209 16:24:19.750904 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5sc2\" (UniqueName: \"kubernetes.io/projected/74ee2091-d89d-4945-af03-5b9887bdfa24-kube-api-access-x5sc2\") pod \"certified-operators-lnlst\" (UID: \"74ee2091-d89d-4945-af03-5b9887bdfa24\") " pod="openshift-marketplace/certified-operators-lnlst" Dec 09 16:24:19 crc kubenswrapper[4894]: I1209 16:24:19.750977 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74ee2091-d89d-4945-af03-5b9887bdfa24-catalog-content\") pod \"certified-operators-lnlst\" (UID: \"74ee2091-d89d-4945-af03-5b9887bdfa24\") " pod="openshift-marketplace/certified-operators-lnlst" Dec 09 16:24:19 crc kubenswrapper[4894]: I1209 16:24:19.751425 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74ee2091-d89d-4945-af03-5b9887bdfa24-catalog-content\") pod \"certified-operators-lnlst\" (UID: \"74ee2091-d89d-4945-af03-5b9887bdfa24\") " pod="openshift-marketplace/certified-operators-lnlst" Dec 09 16:24:19 crc kubenswrapper[4894]: I1209 16:24:19.751443 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74ee2091-d89d-4945-af03-5b9887bdfa24-utilities\") pod \"certified-operators-lnlst\" (UID: \"74ee2091-d89d-4945-af03-5b9887bdfa24\") " pod="openshift-marketplace/certified-operators-lnlst" Dec 09 16:24:19 crc kubenswrapper[4894]: I1209 16:24:19.775109 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5sc2\" (UniqueName: \"kubernetes.io/projected/74ee2091-d89d-4945-af03-5b9887bdfa24-kube-api-access-x5sc2\") pod \"certified-operators-lnlst\" (UID: \"74ee2091-d89d-4945-af03-5b9887bdfa24\") " pod="openshift-marketplace/certified-operators-lnlst" Dec 09 16:24:19 crc kubenswrapper[4894]: I1209 16:24:19.878578 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lnlst" Dec 09 16:24:20 crc kubenswrapper[4894]: I1209 16:24:20.409755 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lnlst"] Dec 09 16:24:20 crc kubenswrapper[4894]: I1209 16:24:20.821684 4894 generic.go:334] "Generic (PLEG): container finished" podID="74ee2091-d89d-4945-af03-5b9887bdfa24" containerID="f63d61cd8a843613c1d00a8dc44375294a547d79110e11143738101cdd201a65" exitCode=0 Dec 09 16:24:20 crc kubenswrapper[4894]: I1209 16:24:20.821731 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnlst" event={"ID":"74ee2091-d89d-4945-af03-5b9887bdfa24","Type":"ContainerDied","Data":"f63d61cd8a843613c1d00a8dc44375294a547d79110e11143738101cdd201a65"} Dec 09 16:24:20 crc kubenswrapper[4894]: I1209 16:24:20.821764 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnlst" event={"ID":"74ee2091-d89d-4945-af03-5b9887bdfa24","Type":"ContainerStarted","Data":"df416025652fa71db66c630370b5eec8678d35aac24d11431e254bf98a468c91"} Dec 09 16:24:25 crc kubenswrapper[4894]: I1209 16:24:25.882933 4894 generic.go:334] "Generic (PLEG): container finished" podID="74ee2091-d89d-4945-af03-5b9887bdfa24" containerID="5b72cfefd9241bdc06445f6dcf80c5109c59a376b3e91843424f3c83ef361823" exitCode=0 Dec 09 16:24:25 crc kubenswrapper[4894]: I1209 16:24:25.883069 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnlst" event={"ID":"74ee2091-d89d-4945-af03-5b9887bdfa24","Type":"ContainerDied","Data":"5b72cfefd9241bdc06445f6dcf80c5109c59a376b3e91843424f3c83ef361823"} Dec 09 16:24:27 crc kubenswrapper[4894]: I1209 16:24:27.106783 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:24:27 crc kubenswrapper[4894]: E1209 16:24:27.107142 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:24:29 crc kubenswrapper[4894]: I1209 16:24:29.945868 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnlst" event={"ID":"74ee2091-d89d-4945-af03-5b9887bdfa24","Type":"ContainerStarted","Data":"9d0d808aaf5c6a7d109a94a5530688e8d06c0b619317ad4409e3d0248e473339"} Dec 09 16:24:29 crc kubenswrapper[4894]: I1209 16:24:29.984783 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-lnlst" podStartSLOduration=2.9600052569999997 podStartE2EDuration="10.984750802s" podCreationTimestamp="2025-12-09 16:24:19 +0000 UTC" firstStartedPulling="2025-12-09 16:24:20.824001541 +0000 UTC m=+3155.143212220" lastFinishedPulling="2025-12-09 16:24:28.848747066 +0000 UTC m=+3163.167957765" observedRunningTime="2025-12-09 16:24:29.971201146 +0000 UTC m=+3164.290411815" watchObservedRunningTime="2025-12-09 16:24:29.984750802 +0000 UTC m=+3164.303961551" Dec 09 16:24:39 crc kubenswrapper[4894]: I1209 16:24:39.878895 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-lnlst" Dec 09 16:24:39 crc kubenswrapper[4894]: I1209 16:24:39.879609 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-lnlst" Dec 09 16:24:39 crc kubenswrapper[4894]: I1209 16:24:39.944595 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-lnlst" Dec 09 16:24:40 crc kubenswrapper[4894]: I1209 16:24:40.151695 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-lnlst" Dec 09 16:24:40 crc kubenswrapper[4894]: I1209 16:24:40.243724 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-lnlst"] Dec 09 16:24:40 crc kubenswrapper[4894]: I1209 16:24:40.310778 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7rmdd"] Dec 09 16:24:40 crc kubenswrapper[4894]: I1209 16:24:40.311021 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7rmdd" podUID="59467bee-13bb-4c8c-8617-b485ac0d0de0" containerName="registry-server" containerID="cri-o://ac902c38d30b45aa18cdb7c6d5a8c84b7d0a1e1b748c200b20afecfb0ad26847" gracePeriod=2 Dec 09 16:24:40 crc kubenswrapper[4894]: I1209 16:24:40.738938 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7rmdd" Dec 09 16:24:40 crc kubenswrapper[4894]: I1209 16:24:40.852325 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z2kw7\" (UniqueName: \"kubernetes.io/projected/59467bee-13bb-4c8c-8617-b485ac0d0de0-kube-api-access-z2kw7\") pod \"59467bee-13bb-4c8c-8617-b485ac0d0de0\" (UID: \"59467bee-13bb-4c8c-8617-b485ac0d0de0\") " Dec 09 16:24:40 crc kubenswrapper[4894]: I1209 16:24:40.852404 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59467bee-13bb-4c8c-8617-b485ac0d0de0-utilities\") pod \"59467bee-13bb-4c8c-8617-b485ac0d0de0\" (UID: \"59467bee-13bb-4c8c-8617-b485ac0d0de0\") " Dec 09 16:24:40 crc kubenswrapper[4894]: I1209 16:24:40.852597 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59467bee-13bb-4c8c-8617-b485ac0d0de0-catalog-content\") pod \"59467bee-13bb-4c8c-8617-b485ac0d0de0\" (UID: \"59467bee-13bb-4c8c-8617-b485ac0d0de0\") " Dec 09 16:24:40 crc kubenswrapper[4894]: I1209 16:24:40.853264 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59467bee-13bb-4c8c-8617-b485ac0d0de0-utilities" (OuterVolumeSpecName: "utilities") pod "59467bee-13bb-4c8c-8617-b485ac0d0de0" (UID: "59467bee-13bb-4c8c-8617-b485ac0d0de0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:24:40 crc kubenswrapper[4894]: I1209 16:24:40.861964 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59467bee-13bb-4c8c-8617-b485ac0d0de0-kube-api-access-z2kw7" (OuterVolumeSpecName: "kube-api-access-z2kw7") pod "59467bee-13bb-4c8c-8617-b485ac0d0de0" (UID: "59467bee-13bb-4c8c-8617-b485ac0d0de0"). InnerVolumeSpecName "kube-api-access-z2kw7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:24:40 crc kubenswrapper[4894]: I1209 16:24:40.904142 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/59467bee-13bb-4c8c-8617-b485ac0d0de0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "59467bee-13bb-4c8c-8617-b485ac0d0de0" (UID: "59467bee-13bb-4c8c-8617-b485ac0d0de0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:24:40 crc kubenswrapper[4894]: I1209 16:24:40.955441 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z2kw7\" (UniqueName: \"kubernetes.io/projected/59467bee-13bb-4c8c-8617-b485ac0d0de0-kube-api-access-z2kw7\") on node \"crc\" DevicePath \"\"" Dec 09 16:24:40 crc kubenswrapper[4894]: I1209 16:24:40.955712 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/59467bee-13bb-4c8c-8617-b485ac0d0de0-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:24:40 crc kubenswrapper[4894]: I1209 16:24:40.955722 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/59467bee-13bb-4c8c-8617-b485ac0d0de0-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:24:41 crc kubenswrapper[4894]: I1209 16:24:41.083351 4894 generic.go:334] "Generic (PLEG): container finished" podID="59467bee-13bb-4c8c-8617-b485ac0d0de0" containerID="ac902c38d30b45aa18cdb7c6d5a8c84b7d0a1e1b748c200b20afecfb0ad26847" exitCode=0 Dec 09 16:24:41 crc kubenswrapper[4894]: I1209 16:24:41.083396 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7rmdd" Dec 09 16:24:41 crc kubenswrapper[4894]: I1209 16:24:41.083447 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7rmdd" event={"ID":"59467bee-13bb-4c8c-8617-b485ac0d0de0","Type":"ContainerDied","Data":"ac902c38d30b45aa18cdb7c6d5a8c84b7d0a1e1b748c200b20afecfb0ad26847"} Dec 09 16:24:41 crc kubenswrapper[4894]: I1209 16:24:41.083495 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7rmdd" event={"ID":"59467bee-13bb-4c8c-8617-b485ac0d0de0","Type":"ContainerDied","Data":"722a80807a9fbfa646c7e0a173ac9b8bfcf496841908a449acdfe1f451b96f6d"} Dec 09 16:24:41 crc kubenswrapper[4894]: I1209 16:24:41.083517 4894 scope.go:117] "RemoveContainer" containerID="ac902c38d30b45aa18cdb7c6d5a8c84b7d0a1e1b748c200b20afecfb0ad26847" Dec 09 16:24:41 crc kubenswrapper[4894]: I1209 16:24:41.101851 4894 scope.go:117] "RemoveContainer" containerID="dde39435d23a41d472f494e1c9f060fa4ecda4e9d6708ec0b324154d33476512" Dec 09 16:24:41 crc kubenswrapper[4894]: I1209 16:24:41.114762 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7rmdd"] Dec 09 16:24:41 crc kubenswrapper[4894]: I1209 16:24:41.122035 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7rmdd"] Dec 09 16:24:41 crc kubenswrapper[4894]: I1209 16:24:41.140399 4894 scope.go:117] "RemoveContainer" containerID="d4c83d52bd72e094875513bbc2f3405469de0ca7f7ea19d34f7e5882f5afcd8d" Dec 09 16:24:41 crc kubenswrapper[4894]: I1209 16:24:41.171206 4894 scope.go:117] "RemoveContainer" containerID="ac902c38d30b45aa18cdb7c6d5a8c84b7d0a1e1b748c200b20afecfb0ad26847" Dec 09 16:24:41 crc kubenswrapper[4894]: E1209 16:24:41.171674 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ac902c38d30b45aa18cdb7c6d5a8c84b7d0a1e1b748c200b20afecfb0ad26847\": container with ID starting with ac902c38d30b45aa18cdb7c6d5a8c84b7d0a1e1b748c200b20afecfb0ad26847 not found: ID does not exist" containerID="ac902c38d30b45aa18cdb7c6d5a8c84b7d0a1e1b748c200b20afecfb0ad26847" Dec 09 16:24:41 crc kubenswrapper[4894]: I1209 16:24:41.171712 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ac902c38d30b45aa18cdb7c6d5a8c84b7d0a1e1b748c200b20afecfb0ad26847"} err="failed to get container status \"ac902c38d30b45aa18cdb7c6d5a8c84b7d0a1e1b748c200b20afecfb0ad26847\": rpc error: code = NotFound desc = could not find container \"ac902c38d30b45aa18cdb7c6d5a8c84b7d0a1e1b748c200b20afecfb0ad26847\": container with ID starting with ac902c38d30b45aa18cdb7c6d5a8c84b7d0a1e1b748c200b20afecfb0ad26847 not found: ID does not exist" Dec 09 16:24:41 crc kubenswrapper[4894]: I1209 16:24:41.171741 4894 scope.go:117] "RemoveContainer" containerID="dde39435d23a41d472f494e1c9f060fa4ecda4e9d6708ec0b324154d33476512" Dec 09 16:24:41 crc kubenswrapper[4894]: E1209 16:24:41.172150 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dde39435d23a41d472f494e1c9f060fa4ecda4e9d6708ec0b324154d33476512\": container with ID starting with dde39435d23a41d472f494e1c9f060fa4ecda4e9d6708ec0b324154d33476512 not found: ID does not exist" containerID="dde39435d23a41d472f494e1c9f060fa4ecda4e9d6708ec0b324154d33476512" Dec 09 16:24:41 crc kubenswrapper[4894]: I1209 16:24:41.172230 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dde39435d23a41d472f494e1c9f060fa4ecda4e9d6708ec0b324154d33476512"} err="failed to get container status \"dde39435d23a41d472f494e1c9f060fa4ecda4e9d6708ec0b324154d33476512\": rpc error: code = NotFound desc = could not find container \"dde39435d23a41d472f494e1c9f060fa4ecda4e9d6708ec0b324154d33476512\": container with ID starting with dde39435d23a41d472f494e1c9f060fa4ecda4e9d6708ec0b324154d33476512 not found: ID does not exist" Dec 09 16:24:41 crc kubenswrapper[4894]: I1209 16:24:41.172303 4894 scope.go:117] "RemoveContainer" containerID="d4c83d52bd72e094875513bbc2f3405469de0ca7f7ea19d34f7e5882f5afcd8d" Dec 09 16:24:41 crc kubenswrapper[4894]: E1209 16:24:41.172758 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4c83d52bd72e094875513bbc2f3405469de0ca7f7ea19d34f7e5882f5afcd8d\": container with ID starting with d4c83d52bd72e094875513bbc2f3405469de0ca7f7ea19d34f7e5882f5afcd8d not found: ID does not exist" containerID="d4c83d52bd72e094875513bbc2f3405469de0ca7f7ea19d34f7e5882f5afcd8d" Dec 09 16:24:41 crc kubenswrapper[4894]: I1209 16:24:41.172782 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4c83d52bd72e094875513bbc2f3405469de0ca7f7ea19d34f7e5882f5afcd8d"} err="failed to get container status \"d4c83d52bd72e094875513bbc2f3405469de0ca7f7ea19d34f7e5882f5afcd8d\": rpc error: code = NotFound desc = could not find container \"d4c83d52bd72e094875513bbc2f3405469de0ca7f7ea19d34f7e5882f5afcd8d\": container with ID starting with d4c83d52bd72e094875513bbc2f3405469de0ca7f7ea19d34f7e5882f5afcd8d not found: ID does not exist" Dec 09 16:24:42 crc kubenswrapper[4894]: I1209 16:24:42.107462 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:24:42 crc kubenswrapper[4894]: E1209 16:24:42.107725 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:24:42 crc kubenswrapper[4894]: I1209 16:24:42.117888 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="59467bee-13bb-4c8c-8617-b485ac0d0de0" path="/var/lib/kubelet/pods/59467bee-13bb-4c8c-8617-b485ac0d0de0/volumes" Dec 09 16:24:51 crc kubenswrapper[4894]: I1209 16:24:51.551390 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-c4486db64-xg2r2"] Dec 09 16:24:51 crc kubenswrapper[4894]: E1209 16:24:51.552210 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59467bee-13bb-4c8c-8617-b485ac0d0de0" containerName="extract-utilities" Dec 09 16:24:51 crc kubenswrapper[4894]: I1209 16:24:51.552224 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="59467bee-13bb-4c8c-8617-b485ac0d0de0" containerName="extract-utilities" Dec 09 16:24:51 crc kubenswrapper[4894]: E1209 16:24:51.552231 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59467bee-13bb-4c8c-8617-b485ac0d0de0" containerName="registry-server" Dec 09 16:24:51 crc kubenswrapper[4894]: I1209 16:24:51.552237 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="59467bee-13bb-4c8c-8617-b485ac0d0de0" containerName="registry-server" Dec 09 16:24:51 crc kubenswrapper[4894]: E1209 16:24:51.552272 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59467bee-13bb-4c8c-8617-b485ac0d0de0" containerName="extract-content" Dec 09 16:24:51 crc kubenswrapper[4894]: I1209 16:24:51.552280 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="59467bee-13bb-4c8c-8617-b485ac0d0de0" containerName="extract-content" Dec 09 16:24:51 crc kubenswrapper[4894]: I1209 16:24:51.552461 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="59467bee-13bb-4c8c-8617-b485ac0d0de0" containerName="registry-server" Dec 09 16:24:51 crc kubenswrapper[4894]: I1209 16:24:51.554132 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-c4486db64-xg2r2" Dec 09 16:24:51 crc kubenswrapper[4894]: I1209 16:24:51.576209 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-c4486db64-xg2r2"] Dec 09 16:24:51 crc kubenswrapper[4894]: I1209 16:24:51.683798 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fm8zt\" (UniqueName: \"kubernetes.io/projected/6a5bd235-8edd-4264-85f0-fabe63b56012-kube-api-access-fm8zt\") pod \"openstack-operator-controller-operator-c4486db64-xg2r2\" (UID: \"6a5bd235-8edd-4264-85f0-fabe63b56012\") " pod="openstack-operators/openstack-operator-controller-operator-c4486db64-xg2r2" Dec 09 16:24:51 crc kubenswrapper[4894]: I1209 16:24:51.785836 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fm8zt\" (UniqueName: \"kubernetes.io/projected/6a5bd235-8edd-4264-85f0-fabe63b56012-kube-api-access-fm8zt\") pod \"openstack-operator-controller-operator-c4486db64-xg2r2\" (UID: \"6a5bd235-8edd-4264-85f0-fabe63b56012\") " pod="openstack-operators/openstack-operator-controller-operator-c4486db64-xg2r2" Dec 09 16:24:51 crc kubenswrapper[4894]: I1209 16:24:51.808306 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fm8zt\" (UniqueName: \"kubernetes.io/projected/6a5bd235-8edd-4264-85f0-fabe63b56012-kube-api-access-fm8zt\") pod \"openstack-operator-controller-operator-c4486db64-xg2r2\" (UID: \"6a5bd235-8edd-4264-85f0-fabe63b56012\") " pod="openstack-operators/openstack-operator-controller-operator-c4486db64-xg2r2" Dec 09 16:24:51 crc kubenswrapper[4894]: I1209 16:24:51.873886 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-c4486db64-xg2r2" Dec 09 16:24:52 crc kubenswrapper[4894]: I1209 16:24:52.416957 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-c4486db64-xg2r2"] Dec 09 16:24:53 crc kubenswrapper[4894]: I1209 16:24:53.224278 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-c4486db64-xg2r2" event={"ID":"6a5bd235-8edd-4264-85f0-fabe63b56012","Type":"ContainerStarted","Data":"cbaf4d3db41f1da18bd0e9db3c2d8cdbf5fc67e658f909aadfb396e3e49f1e43"} Dec 09 16:24:53 crc kubenswrapper[4894]: I1209 16:24:53.224627 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-c4486db64-xg2r2" event={"ID":"6a5bd235-8edd-4264-85f0-fabe63b56012","Type":"ContainerStarted","Data":"6babb22c4c4dba06910a196a2ede37131eb36773f1a0b89201a429d7d64335d9"} Dec 09 16:24:53 crc kubenswrapper[4894]: I1209 16:24:53.225122 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-c4486db64-xg2r2" Dec 09 16:24:53 crc kubenswrapper[4894]: I1209 16:24:53.272267 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-c4486db64-xg2r2" podStartSLOduration=2.272250169 podStartE2EDuration="2.272250169s" podCreationTimestamp="2025-12-09 16:24:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:24:53.268150288 +0000 UTC m=+3187.587360997" watchObservedRunningTime="2025-12-09 16:24:53.272250169 +0000 UTC m=+3187.591460838" Dec 09 16:24:54 crc kubenswrapper[4894]: I1209 16:24:54.107203 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:24:54 crc kubenswrapper[4894]: E1209 16:24:54.107511 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:25:01 crc kubenswrapper[4894]: I1209 16:25:01.876477 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-c4486db64-xg2r2" Dec 09 16:25:01 crc kubenswrapper[4894]: I1209 16:25:01.963506 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7979d445b4-d7b6w"] Dec 09 16:25:01 crc kubenswrapper[4894]: I1209 16:25:01.963769 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-d7b6w" podUID="463649a6-8336-419d-9017-e62cc33699cd" containerName="operator" containerID="cri-o://cef03d77ca55f37dc7b9f83ff3c40435d4574a4b032a1f9ef8295c9e58edf8b0" gracePeriod=10 Dec 09 16:25:02 crc kubenswrapper[4894]: I1209 16:25:02.317894 4894 generic.go:334] "Generic (PLEG): container finished" podID="463649a6-8336-419d-9017-e62cc33699cd" containerID="cef03d77ca55f37dc7b9f83ff3c40435d4574a4b032a1f9ef8295c9e58edf8b0" exitCode=0 Dec 09 16:25:02 crc kubenswrapper[4894]: I1209 16:25:02.317959 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-d7b6w" event={"ID":"463649a6-8336-419d-9017-e62cc33699cd","Type":"ContainerDied","Data":"cef03d77ca55f37dc7b9f83ff3c40435d4574a4b032a1f9ef8295c9e58edf8b0"} Dec 09 16:25:02 crc kubenswrapper[4894]: I1209 16:25:02.573687 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-d7b6w" Dec 09 16:25:02 crc kubenswrapper[4894]: I1209 16:25:02.672872 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6gblr\" (UniqueName: \"kubernetes.io/projected/463649a6-8336-419d-9017-e62cc33699cd-kube-api-access-6gblr\") pod \"463649a6-8336-419d-9017-e62cc33699cd\" (UID: \"463649a6-8336-419d-9017-e62cc33699cd\") " Dec 09 16:25:02 crc kubenswrapper[4894]: I1209 16:25:02.678488 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/463649a6-8336-419d-9017-e62cc33699cd-kube-api-access-6gblr" (OuterVolumeSpecName: "kube-api-access-6gblr") pod "463649a6-8336-419d-9017-e62cc33699cd" (UID: "463649a6-8336-419d-9017-e62cc33699cd"). InnerVolumeSpecName "kube-api-access-6gblr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:25:02 crc kubenswrapper[4894]: I1209 16:25:02.775205 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6gblr\" (UniqueName: \"kubernetes.io/projected/463649a6-8336-419d-9017-e62cc33699cd-kube-api-access-6gblr\") on node \"crc\" DevicePath \"\"" Dec 09 16:25:03 crc kubenswrapper[4894]: I1209 16:25:03.331321 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-d7b6w" event={"ID":"463649a6-8336-419d-9017-e62cc33699cd","Type":"ContainerDied","Data":"58b2aead28ef3ca128094eee4b51d92fc4402f0c01024dbbc54a90851344229b"} Dec 09 16:25:03 crc kubenswrapper[4894]: I1209 16:25:03.331387 4894 scope.go:117] "RemoveContainer" containerID="cef03d77ca55f37dc7b9f83ff3c40435d4574a4b032a1f9ef8295c9e58edf8b0" Dec 09 16:25:03 crc kubenswrapper[4894]: I1209 16:25:03.331422 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-7979d445b4-d7b6w" Dec 09 16:25:03 crc kubenswrapper[4894]: I1209 16:25:03.389454 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7979d445b4-d7b6w"] Dec 09 16:25:03 crc kubenswrapper[4894]: I1209 16:25:03.397068 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-7979d445b4-d7b6w"] Dec 09 16:25:04 crc kubenswrapper[4894]: I1209 16:25:04.121152 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="463649a6-8336-419d-9017-e62cc33699cd" path="/var/lib/kubelet/pods/463649a6-8336-419d-9017-e62cc33699cd/volumes" Dec 09 16:25:05 crc kubenswrapper[4894]: I1209 16:25:05.107316 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:25:05 crc kubenswrapper[4894]: E1209 16:25:05.108149 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:25:16 crc kubenswrapper[4894]: I1209 16:25:16.113406 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:25:16 crc kubenswrapper[4894]: E1209 16:25:16.114211 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:25:28 crc kubenswrapper[4894]: I1209 16:25:28.106909 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:25:28 crc kubenswrapper[4894]: E1209 16:25:28.109472 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:25:34 crc kubenswrapper[4894]: I1209 16:25:34.587473 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-6dc759c869-b8lkg"] Dec 09 16:25:34 crc kubenswrapper[4894]: E1209 16:25:34.592630 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="463649a6-8336-419d-9017-e62cc33699cd" containerName="operator" Dec 09 16:25:34 crc kubenswrapper[4894]: I1209 16:25:34.592687 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="463649a6-8336-419d-9017-e62cc33699cd" containerName="operator" Dec 09 16:25:34 crc kubenswrapper[4894]: I1209 16:25:34.593297 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="463649a6-8336-419d-9017-e62cc33699cd" containerName="operator" Dec 09 16:25:34 crc kubenswrapper[4894]: I1209 16:25:34.594970 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-6dc759c869-b8lkg" Dec 09 16:25:34 crc kubenswrapper[4894]: I1209 16:25:34.601540 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-6dc759c869-b8lkg"] Dec 09 16:25:34 crc kubenswrapper[4894]: I1209 16:25:34.715384 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m849r\" (UniqueName: \"kubernetes.io/projected/23370728-b91c-4d67-9c3b-a61aef727c25-kube-api-access-m849r\") pod \"test-operator-controller-manager-6dc759c869-b8lkg\" (UID: \"23370728-b91c-4d67-9c3b-a61aef727c25\") " pod="openstack-operators/test-operator-controller-manager-6dc759c869-b8lkg" Dec 09 16:25:34 crc kubenswrapper[4894]: I1209 16:25:34.818072 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m849r\" (UniqueName: \"kubernetes.io/projected/23370728-b91c-4d67-9c3b-a61aef727c25-kube-api-access-m849r\") pod \"test-operator-controller-manager-6dc759c869-b8lkg\" (UID: \"23370728-b91c-4d67-9c3b-a61aef727c25\") " pod="openstack-operators/test-operator-controller-manager-6dc759c869-b8lkg" Dec 09 16:25:34 crc kubenswrapper[4894]: I1209 16:25:34.847894 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m849r\" (UniqueName: \"kubernetes.io/projected/23370728-b91c-4d67-9c3b-a61aef727c25-kube-api-access-m849r\") pod \"test-operator-controller-manager-6dc759c869-b8lkg\" (UID: \"23370728-b91c-4d67-9c3b-a61aef727c25\") " pod="openstack-operators/test-operator-controller-manager-6dc759c869-b8lkg" Dec 09 16:25:34 crc kubenswrapper[4894]: I1209 16:25:34.938211 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-6dc759c869-b8lkg" Dec 09 16:25:35 crc kubenswrapper[4894]: I1209 16:25:35.414125 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-6dc759c869-b8lkg"] Dec 09 16:25:35 crc kubenswrapper[4894]: W1209 16:25:35.416689 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23370728_b91c_4d67_9c3b_a61aef727c25.slice/crio-0a63dc1cc27f6896813f849dfc3821c6f6db598d0c3305220b97b1bb80bff2a5 WatchSource:0}: Error finding container 0a63dc1cc27f6896813f849dfc3821c6f6db598d0c3305220b97b1bb80bff2a5: Status 404 returned error can't find the container with id 0a63dc1cc27f6896813f849dfc3821c6f6db598d0c3305220b97b1bb80bff2a5 Dec 09 16:25:35 crc kubenswrapper[4894]: I1209 16:25:35.418764 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 16:25:35 crc kubenswrapper[4894]: I1209 16:25:35.706611 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6dc759c869-b8lkg" event={"ID":"23370728-b91c-4d67-9c3b-a61aef727c25","Type":"ContainerStarted","Data":"0a63dc1cc27f6896813f849dfc3821c6f6db598d0c3305220b97b1bb80bff2a5"} Dec 09 16:25:36 crc kubenswrapper[4894]: I1209 16:25:36.716622 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6dc759c869-b8lkg" event={"ID":"23370728-b91c-4d67-9c3b-a61aef727c25","Type":"ContainerStarted","Data":"d1134afc550bd8745ac8c47206e952de9a35a4c0a7af3859322ed44638cddd9d"} Dec 09 16:25:37 crc kubenswrapper[4894]: I1209 16:25:37.729969 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-6dc759c869-b8lkg" event={"ID":"23370728-b91c-4d67-9c3b-a61aef727c25","Type":"ContainerStarted","Data":"963671ce239a3f5cd6c30a6098828393ce4ce550481d58b8bf1b14f61485d040"} Dec 09 16:25:37 crc kubenswrapper[4894]: I1209 16:25:37.731564 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-6dc759c869-b8lkg" Dec 09 16:25:37 crc kubenswrapper[4894]: I1209 16:25:37.765967 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-6dc759c869-b8lkg" podStartSLOduration=2.747914729 podStartE2EDuration="3.765939348s" podCreationTimestamp="2025-12-09 16:25:34 +0000 UTC" firstStartedPulling="2025-12-09 16:25:35.41851606 +0000 UTC m=+3229.737726729" lastFinishedPulling="2025-12-09 16:25:36.436540679 +0000 UTC m=+3230.755751348" observedRunningTime="2025-12-09 16:25:37.748779355 +0000 UTC m=+3232.067990034" watchObservedRunningTime="2025-12-09 16:25:37.765939348 +0000 UTC m=+3232.085150037" Dec 09 16:25:41 crc kubenswrapper[4894]: I1209 16:25:41.106588 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:25:41 crc kubenswrapper[4894]: E1209 16:25:41.107447 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:25:44 crc kubenswrapper[4894]: I1209 16:25:44.942436 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-6dc759c869-b8lkg" Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.030770 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8"] Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.031094 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8" podUID="93107d61-a76c-4ccf-90c3-4277ac94abf0" containerName="manager" containerID="cri-o://3e4d7bf3c4557c742c5380fe554f925744aef6d53f7e1bebe4a47f5e0d3390d2" gracePeriod=10 Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.031624 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8" podUID="93107d61-a76c-4ccf-90c3-4277ac94abf0" containerName="kube-rbac-proxy" containerID="cri-o://d5cb7a833eb3803d582a2aece60dd5f9df69dbbbc23ef667e3422c98f7f4d8b3" gracePeriod=10 Dec 09 16:25:45 crc kubenswrapper[4894]: E1209 16:25:45.222877 4894 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93107d61_a76c_4ccf_90c3_4277ac94abf0.slice/crio-d5cb7a833eb3803d582a2aece60dd5f9df69dbbbc23ef667e3422c98f7f4d8b3.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod93107d61_a76c_4ccf_90c3_4277ac94abf0.slice/crio-3e4d7bf3c4557c742c5380fe554f925744aef6d53f7e1bebe4a47f5e0d3390d2.scope\": RecentStats: unable to find data in memory cache]" Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.485313 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8" Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.558120 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6crfq\" (UniqueName: \"kubernetes.io/projected/93107d61-a76c-4ccf-90c3-4277ac94abf0-kube-api-access-6crfq\") pod \"93107d61-a76c-4ccf-90c3-4277ac94abf0\" (UID: \"93107d61-a76c-4ccf-90c3-4277ac94abf0\") " Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.565260 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/93107d61-a76c-4ccf-90c3-4277ac94abf0-kube-api-access-6crfq" (OuterVolumeSpecName: "kube-api-access-6crfq") pod "93107d61-a76c-4ccf-90c3-4277ac94abf0" (UID: "93107d61-a76c-4ccf-90c3-4277ac94abf0"). InnerVolumeSpecName "kube-api-access-6crfq". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.660115 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6crfq\" (UniqueName: \"kubernetes.io/projected/93107d61-a76c-4ccf-90c3-4277ac94abf0-kube-api-access-6crfq\") on node \"crc\" DevicePath \"\"" Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.813568 4894 generic.go:334] "Generic (PLEG): container finished" podID="93107d61-a76c-4ccf-90c3-4277ac94abf0" containerID="d5cb7a833eb3803d582a2aece60dd5f9df69dbbbc23ef667e3422c98f7f4d8b3" exitCode=0 Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.813608 4894 generic.go:334] "Generic (PLEG): container finished" podID="93107d61-a76c-4ccf-90c3-4277ac94abf0" containerID="3e4d7bf3c4557c742c5380fe554f925744aef6d53f7e1bebe4a47f5e0d3390d2" exitCode=0 Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.813670 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8" event={"ID":"93107d61-a76c-4ccf-90c3-4277ac94abf0","Type":"ContainerDied","Data":"d5cb7a833eb3803d582a2aece60dd5f9df69dbbbc23ef667e3422c98f7f4d8b3"} Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.813698 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8" Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.813723 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8" event={"ID":"93107d61-a76c-4ccf-90c3-4277ac94abf0","Type":"ContainerDied","Data":"3e4d7bf3c4557c742c5380fe554f925744aef6d53f7e1bebe4a47f5e0d3390d2"} Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.813746 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8" event={"ID":"93107d61-a76c-4ccf-90c3-4277ac94abf0","Type":"ContainerDied","Data":"561407eb4fbb557daf9edd331bec86db91ee822f86ca8ac7e8a1041ccf7fbaae"} Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.813772 4894 scope.go:117] "RemoveContainer" containerID="d5cb7a833eb3803d582a2aece60dd5f9df69dbbbc23ef667e3422c98f7f4d8b3" Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.845036 4894 scope.go:117] "RemoveContainer" containerID="3e4d7bf3c4557c742c5380fe554f925744aef6d53f7e1bebe4a47f5e0d3390d2" Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.885594 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8"] Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.894438 4894 scope.go:117] "RemoveContainer" containerID="d5cb7a833eb3803d582a2aece60dd5f9df69dbbbc23ef667e3422c98f7f4d8b3" Dec 09 16:25:45 crc kubenswrapper[4894]: E1209 16:25:45.895030 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5cb7a833eb3803d582a2aece60dd5f9df69dbbbc23ef667e3422c98f7f4d8b3\": container with ID starting with d5cb7a833eb3803d582a2aece60dd5f9df69dbbbc23ef667e3422c98f7f4d8b3 not found: ID does not exist" containerID="d5cb7a833eb3803d582a2aece60dd5f9df69dbbbc23ef667e3422c98f7f4d8b3" Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.895121 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5cb7a833eb3803d582a2aece60dd5f9df69dbbbc23ef667e3422c98f7f4d8b3"} err="failed to get container status \"d5cb7a833eb3803d582a2aece60dd5f9df69dbbbc23ef667e3422c98f7f4d8b3\": rpc error: code = NotFound desc = could not find container \"d5cb7a833eb3803d582a2aece60dd5f9df69dbbbc23ef667e3422c98f7f4d8b3\": container with ID starting with d5cb7a833eb3803d582a2aece60dd5f9df69dbbbc23ef667e3422c98f7f4d8b3 not found: ID does not exist" Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.895166 4894 scope.go:117] "RemoveContainer" containerID="3e4d7bf3c4557c742c5380fe554f925744aef6d53f7e1bebe4a47f5e0d3390d2" Dec 09 16:25:45 crc kubenswrapper[4894]: E1209 16:25:45.895629 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3e4d7bf3c4557c742c5380fe554f925744aef6d53f7e1bebe4a47f5e0d3390d2\": container with ID starting with 3e4d7bf3c4557c742c5380fe554f925744aef6d53f7e1bebe4a47f5e0d3390d2 not found: ID does not exist" containerID="3e4d7bf3c4557c742c5380fe554f925744aef6d53f7e1bebe4a47f5e0d3390d2" Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.895718 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e4d7bf3c4557c742c5380fe554f925744aef6d53f7e1bebe4a47f5e0d3390d2"} err="failed to get container status \"3e4d7bf3c4557c742c5380fe554f925744aef6d53f7e1bebe4a47f5e0d3390d2\": rpc error: code = NotFound desc = could not find container \"3e4d7bf3c4557c742c5380fe554f925744aef6d53f7e1bebe4a47f5e0d3390d2\": container with ID starting with 3e4d7bf3c4557c742c5380fe554f925744aef6d53f7e1bebe4a47f5e0d3390d2 not found: ID does not exist" Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.895759 4894 scope.go:117] "RemoveContainer" containerID="d5cb7a833eb3803d582a2aece60dd5f9df69dbbbc23ef667e3422c98f7f4d8b3" Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.896132 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5cb7a833eb3803d582a2aece60dd5f9df69dbbbc23ef667e3422c98f7f4d8b3"} err="failed to get container status \"d5cb7a833eb3803d582a2aece60dd5f9df69dbbbc23ef667e3422c98f7f4d8b3\": rpc error: code = NotFound desc = could not find container \"d5cb7a833eb3803d582a2aece60dd5f9df69dbbbc23ef667e3422c98f7f4d8b3\": container with ID starting with d5cb7a833eb3803d582a2aece60dd5f9df69dbbbc23ef667e3422c98f7f4d8b3 not found: ID does not exist" Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.896168 4894 scope.go:117] "RemoveContainer" containerID="3e4d7bf3c4557c742c5380fe554f925744aef6d53f7e1bebe4a47f5e0d3390d2" Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.900846 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3e4d7bf3c4557c742c5380fe554f925744aef6d53f7e1bebe4a47f5e0d3390d2"} err="failed to get container status \"3e4d7bf3c4557c742c5380fe554f925744aef6d53f7e1bebe4a47f5e0d3390d2\": rpc error: code = NotFound desc = could not find container \"3e4d7bf3c4557c742c5380fe554f925744aef6d53f7e1bebe4a47f5e0d3390d2\": container with ID starting with 3e4d7bf3c4557c742c5380fe554f925744aef6d53f7e1bebe4a47f5e0d3390d2 not found: ID does not exist" Dec 09 16:25:45 crc kubenswrapper[4894]: I1209 16:25:45.904263 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/test-operator-controller-manager-5854674fcc-qtsl8"] Dec 09 16:25:46 crc kubenswrapper[4894]: I1209 16:25:46.125497 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="93107d61-a76c-4ccf-90c3-4277ac94abf0" path="/var/lib/kubelet/pods/93107d61-a76c-4ccf-90c3-4277ac94abf0/volumes" Dec 09 16:25:52 crc kubenswrapper[4894]: I1209 16:25:52.106630 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:25:52 crc kubenswrapper[4894]: I1209 16:25:52.891815 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerStarted","Data":"92d4a03025122ef2af3ae2a9840aaf815cef6ed1352687e1bda2ee74c9f07415"} Dec 09 16:27:18 crc kubenswrapper[4894]: I1209 16:27:18.635539 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h4bm2"] Dec 09 16:27:18 crc kubenswrapper[4894]: E1209 16:27:18.636786 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93107d61-a76c-4ccf-90c3-4277ac94abf0" containerName="manager" Dec 09 16:27:18 crc kubenswrapper[4894]: I1209 16:27:18.636808 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="93107d61-a76c-4ccf-90c3-4277ac94abf0" containerName="manager" Dec 09 16:27:18 crc kubenswrapper[4894]: E1209 16:27:18.636837 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="93107d61-a76c-4ccf-90c3-4277ac94abf0" containerName="kube-rbac-proxy" Dec 09 16:27:18 crc kubenswrapper[4894]: I1209 16:27:18.636848 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="93107d61-a76c-4ccf-90c3-4277ac94abf0" containerName="kube-rbac-proxy" Dec 09 16:27:18 crc kubenswrapper[4894]: I1209 16:27:18.637116 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="93107d61-a76c-4ccf-90c3-4277ac94abf0" containerName="kube-rbac-proxy" Dec 09 16:27:18 crc kubenswrapper[4894]: I1209 16:27:18.637154 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="93107d61-a76c-4ccf-90c3-4277ac94abf0" containerName="manager" Dec 09 16:27:18 crc kubenswrapper[4894]: I1209 16:27:18.639318 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h4bm2" Dec 09 16:27:18 crc kubenswrapper[4894]: I1209 16:27:18.673978 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h4bm2"] Dec 09 16:27:18 crc kubenswrapper[4894]: I1209 16:27:18.710661 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8448fa91-d4e8-4331-bbe3-a44f64cb1a01-utilities\") pod \"redhat-operators-h4bm2\" (UID: \"8448fa91-d4e8-4331-bbe3-a44f64cb1a01\") " pod="openshift-marketplace/redhat-operators-h4bm2" Dec 09 16:27:18 crc kubenswrapper[4894]: I1209 16:27:18.710809 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-56shc\" (UniqueName: \"kubernetes.io/projected/8448fa91-d4e8-4331-bbe3-a44f64cb1a01-kube-api-access-56shc\") pod \"redhat-operators-h4bm2\" (UID: \"8448fa91-d4e8-4331-bbe3-a44f64cb1a01\") " pod="openshift-marketplace/redhat-operators-h4bm2" Dec 09 16:27:18 crc kubenswrapper[4894]: I1209 16:27:18.710880 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8448fa91-d4e8-4331-bbe3-a44f64cb1a01-catalog-content\") pod \"redhat-operators-h4bm2\" (UID: \"8448fa91-d4e8-4331-bbe3-a44f64cb1a01\") " pod="openshift-marketplace/redhat-operators-h4bm2" Dec 09 16:27:18 crc kubenswrapper[4894]: I1209 16:27:18.812908 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-56shc\" (UniqueName: \"kubernetes.io/projected/8448fa91-d4e8-4331-bbe3-a44f64cb1a01-kube-api-access-56shc\") pod \"redhat-operators-h4bm2\" (UID: \"8448fa91-d4e8-4331-bbe3-a44f64cb1a01\") " pod="openshift-marketplace/redhat-operators-h4bm2" Dec 09 16:27:18 crc kubenswrapper[4894]: I1209 16:27:18.813068 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8448fa91-d4e8-4331-bbe3-a44f64cb1a01-catalog-content\") pod \"redhat-operators-h4bm2\" (UID: \"8448fa91-d4e8-4331-bbe3-a44f64cb1a01\") " pod="openshift-marketplace/redhat-operators-h4bm2" Dec 09 16:27:18 crc kubenswrapper[4894]: I1209 16:27:18.813134 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8448fa91-d4e8-4331-bbe3-a44f64cb1a01-utilities\") pod \"redhat-operators-h4bm2\" (UID: \"8448fa91-d4e8-4331-bbe3-a44f64cb1a01\") " pod="openshift-marketplace/redhat-operators-h4bm2" Dec 09 16:27:18 crc kubenswrapper[4894]: I1209 16:27:18.813788 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8448fa91-d4e8-4331-bbe3-a44f64cb1a01-catalog-content\") pod \"redhat-operators-h4bm2\" (UID: \"8448fa91-d4e8-4331-bbe3-a44f64cb1a01\") " pod="openshift-marketplace/redhat-operators-h4bm2" Dec 09 16:27:18 crc kubenswrapper[4894]: I1209 16:27:18.813960 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8448fa91-d4e8-4331-bbe3-a44f64cb1a01-utilities\") pod \"redhat-operators-h4bm2\" (UID: \"8448fa91-d4e8-4331-bbe3-a44f64cb1a01\") " pod="openshift-marketplace/redhat-operators-h4bm2" Dec 09 16:27:18 crc kubenswrapper[4894]: I1209 16:27:18.847544 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-56shc\" (UniqueName: \"kubernetes.io/projected/8448fa91-d4e8-4331-bbe3-a44f64cb1a01-kube-api-access-56shc\") pod \"redhat-operators-h4bm2\" (UID: \"8448fa91-d4e8-4331-bbe3-a44f64cb1a01\") " pod="openshift-marketplace/redhat-operators-h4bm2" Dec 09 16:27:18 crc kubenswrapper[4894]: I1209 16:27:18.969526 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h4bm2" Dec 09 16:27:19 crc kubenswrapper[4894]: I1209 16:27:19.443242 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h4bm2"] Dec 09 16:27:19 crc kubenswrapper[4894]: I1209 16:27:19.807788 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h4bm2" event={"ID":"8448fa91-d4e8-4331-bbe3-a44f64cb1a01","Type":"ContainerStarted","Data":"d4ce7e8cea22fd5488e76c539b9a879c63073f67de00609b5a48ef718c6b13db"} Dec 09 16:27:19 crc kubenswrapper[4894]: I1209 16:27:19.808030 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h4bm2" event={"ID":"8448fa91-d4e8-4331-bbe3-a44f64cb1a01","Type":"ContainerStarted","Data":"11096ffd6574a3263287615bf9499a40ffce890bc18ab568f9994e36dc79a624"} Dec 09 16:27:20 crc kubenswrapper[4894]: I1209 16:27:20.817203 4894 generic.go:334] "Generic (PLEG): container finished" podID="8448fa91-d4e8-4331-bbe3-a44f64cb1a01" containerID="d4ce7e8cea22fd5488e76c539b9a879c63073f67de00609b5a48ef718c6b13db" exitCode=0 Dec 09 16:27:20 crc kubenswrapper[4894]: I1209 16:27:20.817337 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h4bm2" event={"ID":"8448fa91-d4e8-4331-bbe3-a44f64cb1a01","Type":"ContainerDied","Data":"d4ce7e8cea22fd5488e76c539b9a879c63073f67de00609b5a48ef718c6b13db"} Dec 09 16:27:22 crc kubenswrapper[4894]: I1209 16:27:22.840612 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h4bm2" event={"ID":"8448fa91-d4e8-4331-bbe3-a44f64cb1a01","Type":"ContainerStarted","Data":"440116e5bf480658d386c6f65f881a19dbd0555ddae038c0771397aedd137173"} Dec 09 16:27:27 crc kubenswrapper[4894]: I1209 16:27:27.905147 4894 generic.go:334] "Generic (PLEG): container finished" podID="8448fa91-d4e8-4331-bbe3-a44f64cb1a01" containerID="440116e5bf480658d386c6f65f881a19dbd0555ddae038c0771397aedd137173" exitCode=0 Dec 09 16:27:27 crc kubenswrapper[4894]: I1209 16:27:27.905751 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h4bm2" event={"ID":"8448fa91-d4e8-4331-bbe3-a44f64cb1a01","Type":"ContainerDied","Data":"440116e5bf480658d386c6f65f881a19dbd0555ddae038c0771397aedd137173"} Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.361046 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest-s00-full"] Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.375723 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.383213 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.383547 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-fxksq" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.383566 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.383768 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s0" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.403256 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s00-full"] Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.454143 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/675a2b9e-a4fe-4ced-9997-c82358f8550c-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.454271 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/675a2b9e-a4fe-4ced-9997-c82358f8550c-config-data\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.454312 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/675a2b9e-a4fe-4ced-9997-c82358f8550c-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.454343 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-ceph\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.454385 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/675a2b9e-a4fe-4ced-9997-c82358f8550c-openstack-config\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.454413 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-openstack-config-secret\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.454486 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.454516 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-ssh-key\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.454558 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-ca-certs\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.454630 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5bm79\" (UniqueName: \"kubernetes.io/projected/675a2b9e-a4fe-4ced-9997-c82358f8550c-kube-api-access-5bm79\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.556595 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.556887 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-ssh-key\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.556938 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-ca-certs\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.557010 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5bm79\" (UniqueName: \"kubernetes.io/projected/675a2b9e-a4fe-4ced-9997-c82358f8550c-kube-api-access-5bm79\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.557017 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.557112 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/675a2b9e-a4fe-4ced-9997-c82358f8550c-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.557157 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/675a2b9e-a4fe-4ced-9997-c82358f8550c-config-data\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.557180 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/675a2b9e-a4fe-4ced-9997-c82358f8550c-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.557200 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-ceph\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.557231 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/675a2b9e-a4fe-4ced-9997-c82358f8550c-openstack-config\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.557247 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-openstack-config-secret\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.557917 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/675a2b9e-a4fe-4ced-9997-c82358f8550c-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.558704 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/675a2b9e-a4fe-4ced-9997-c82358f8550c-openstack-config\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.559018 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/675a2b9e-a4fe-4ced-9997-c82358f8550c-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.559131 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/675a2b9e-a4fe-4ced-9997-c82358f8550c-config-data\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.565465 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-ssh-key\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.567218 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-ceph\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.573668 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-openstack-config-secret\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.579616 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-ca-certs\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.580998 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5bm79\" (UniqueName: \"kubernetes.io/projected/675a2b9e-a4fe-4ced-9997-c82358f8550c-kube-api-access-5bm79\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.592136 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s00-full\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.705433 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.919569 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h4bm2" event={"ID":"8448fa91-d4e8-4331-bbe3-a44f64cb1a01","Type":"ContainerStarted","Data":"ab33013bbace00bb0427d3ac0d23af82ae9561e689b2f75570827dbfd3bf94d1"} Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.949745 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h4bm2" podStartSLOduration=3.3956165609999998 podStartE2EDuration="10.949724773s" podCreationTimestamp="2025-12-09 16:27:18 +0000 UTC" firstStartedPulling="2025-12-09 16:27:20.82027292 +0000 UTC m=+3335.139483619" lastFinishedPulling="2025-12-09 16:27:28.374381152 +0000 UTC m=+3342.693591831" observedRunningTime="2025-12-09 16:27:28.936810125 +0000 UTC m=+3343.256020794" watchObservedRunningTime="2025-12-09 16:27:28.949724773 +0000 UTC m=+3343.268935442" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.969882 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h4bm2" Dec 09 16:27:28 crc kubenswrapper[4894]: I1209 16:27:28.969934 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h4bm2" Dec 09 16:27:29 crc kubenswrapper[4894]: I1209 16:27:29.245117 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s00-full"] Dec 09 16:27:29 crc kubenswrapper[4894]: W1209 16:27:29.248163 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod675a2b9e_a4fe_4ced_9997_c82358f8550c.slice/crio-6df7e934fdfad472b46e55e1e2592ac3e1dd73434babb8fb1d78c184226db643 WatchSource:0}: Error finding container 6df7e934fdfad472b46e55e1e2592ac3e1dd73434babb8fb1d78c184226db643: Status 404 returned error can't find the container with id 6df7e934fdfad472b46e55e1e2592ac3e1dd73434babb8fb1d78c184226db643 Dec 09 16:27:29 crc kubenswrapper[4894]: I1209 16:27:29.929951 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"675a2b9e-a4fe-4ced-9997-c82358f8550c","Type":"ContainerStarted","Data":"6df7e934fdfad472b46e55e1e2592ac3e1dd73434babb8fb1d78c184226db643"} Dec 09 16:27:30 crc kubenswrapper[4894]: I1209 16:27:30.013114 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-h4bm2" podUID="8448fa91-d4e8-4331-bbe3-a44f64cb1a01" containerName="registry-server" probeResult="failure" output=< Dec 09 16:27:30 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 09 16:27:30 crc kubenswrapper[4894]: > Dec 09 16:27:40 crc kubenswrapper[4894]: I1209 16:27:40.023862 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-h4bm2" podUID="8448fa91-d4e8-4331-bbe3-a44f64cb1a01" containerName="registry-server" probeResult="failure" output=< Dec 09 16:27:40 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 09 16:27:40 crc kubenswrapper[4894]: > Dec 09 16:27:49 crc kubenswrapper[4894]: I1209 16:27:49.049407 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h4bm2" Dec 09 16:27:49 crc kubenswrapper[4894]: I1209 16:27:49.101530 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h4bm2" Dec 09 16:27:49 crc kubenswrapper[4894]: I1209 16:27:49.843791 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h4bm2"] Dec 09 16:27:50 crc kubenswrapper[4894]: I1209 16:27:50.131810 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-h4bm2" podUID="8448fa91-d4e8-4331-bbe3-a44f64cb1a01" containerName="registry-server" containerID="cri-o://ab33013bbace00bb0427d3ac0d23af82ae9561e689b2f75570827dbfd3bf94d1" gracePeriod=2 Dec 09 16:27:51 crc kubenswrapper[4894]: I1209 16:27:51.146724 4894 generic.go:334] "Generic (PLEG): container finished" podID="8448fa91-d4e8-4331-bbe3-a44f64cb1a01" containerID="ab33013bbace00bb0427d3ac0d23af82ae9561e689b2f75570827dbfd3bf94d1" exitCode=0 Dec 09 16:27:51 crc kubenswrapper[4894]: I1209 16:27:51.146815 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h4bm2" event={"ID":"8448fa91-d4e8-4331-bbe3-a44f64cb1a01","Type":"ContainerDied","Data":"ab33013bbace00bb0427d3ac0d23af82ae9561e689b2f75570827dbfd3bf94d1"} Dec 09 16:27:58 crc kubenswrapper[4894]: E1209 16:27:58.971143 4894 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ab33013bbace00bb0427d3ac0d23af82ae9561e689b2f75570827dbfd3bf94d1 is running failed: container process not found" containerID="ab33013bbace00bb0427d3ac0d23af82ae9561e689b2f75570827dbfd3bf94d1" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 16:27:58 crc kubenswrapper[4894]: E1209 16:27:58.972454 4894 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ab33013bbace00bb0427d3ac0d23af82ae9561e689b2f75570827dbfd3bf94d1 is running failed: container process not found" containerID="ab33013bbace00bb0427d3ac0d23af82ae9561e689b2f75570827dbfd3bf94d1" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 16:27:58 crc kubenswrapper[4894]: E1209 16:27:58.972886 4894 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ab33013bbace00bb0427d3ac0d23af82ae9561e689b2f75570827dbfd3bf94d1 is running failed: container process not found" containerID="ab33013bbace00bb0427d3ac0d23af82ae9561e689b2f75570827dbfd3bf94d1" cmd=["grpc_health_probe","-addr=:50051"] Dec 09 16:27:58 crc kubenswrapper[4894]: E1209 16:27:58.972946 4894 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of ab33013bbace00bb0427d3ac0d23af82ae9561e689b2f75570827dbfd3bf94d1 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-operators-h4bm2" podUID="8448fa91-d4e8-4331-bbe3-a44f64cb1a01" containerName="registry-server" Dec 09 16:28:02 crc kubenswrapper[4894]: E1209 16:28:02.272123 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified" Dec 09 16:28:02 crc kubenswrapper[4894]: E1209 16:28:02.272790 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tempest-tests-tempest-tests-runner,Image:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:false,MountPath:/etc/test_operator,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tempest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tempest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/tempest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ssh-key,ReadOnly:false,MountPath:/var/lib/tempest/id_ecdsa,SubPath:ssh_key,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5bm79,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42480,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42480,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-custom-data-s0,},Optional:nil,},SecretRef:nil,},EnvFromSource{Prefix:,ConfigMapRef:&ConfigMapEnvSource{LocalObjectReference:LocalObjectReference{Name:tempest-tests-tempest-env-vars-s0,},Optional:nil,},SecretRef:nil,},},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tempest-tests-tempest-s00-full_openstack(675a2b9e-a4fe-4ced-9997-c82358f8550c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 16:28:02 crc kubenswrapper[4894]: E1209 16:28:02.274129 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tempest-tests-tempest-s00-full" podUID="675a2b9e-a4fe-4ced-9997-c82358f8550c" Dec 09 16:28:02 crc kubenswrapper[4894]: I1209 16:28:02.761772 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h4bm2" Dec 09 16:28:02 crc kubenswrapper[4894]: I1209 16:28:02.917819 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-56shc\" (UniqueName: \"kubernetes.io/projected/8448fa91-d4e8-4331-bbe3-a44f64cb1a01-kube-api-access-56shc\") pod \"8448fa91-d4e8-4331-bbe3-a44f64cb1a01\" (UID: \"8448fa91-d4e8-4331-bbe3-a44f64cb1a01\") " Dec 09 16:28:02 crc kubenswrapper[4894]: I1209 16:28:02.918342 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8448fa91-d4e8-4331-bbe3-a44f64cb1a01-utilities\") pod \"8448fa91-d4e8-4331-bbe3-a44f64cb1a01\" (UID: \"8448fa91-d4e8-4331-bbe3-a44f64cb1a01\") " Dec 09 16:28:02 crc kubenswrapper[4894]: I1209 16:28:02.918520 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8448fa91-d4e8-4331-bbe3-a44f64cb1a01-catalog-content\") pod \"8448fa91-d4e8-4331-bbe3-a44f64cb1a01\" (UID: \"8448fa91-d4e8-4331-bbe3-a44f64cb1a01\") " Dec 09 16:28:02 crc kubenswrapper[4894]: I1209 16:28:02.919192 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8448fa91-d4e8-4331-bbe3-a44f64cb1a01-utilities" (OuterVolumeSpecName: "utilities") pod "8448fa91-d4e8-4331-bbe3-a44f64cb1a01" (UID: "8448fa91-d4e8-4331-bbe3-a44f64cb1a01"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:28:02 crc kubenswrapper[4894]: I1209 16:28:02.937253 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8448fa91-d4e8-4331-bbe3-a44f64cb1a01-kube-api-access-56shc" (OuterVolumeSpecName: "kube-api-access-56shc") pod "8448fa91-d4e8-4331-bbe3-a44f64cb1a01" (UID: "8448fa91-d4e8-4331-bbe3-a44f64cb1a01"). InnerVolumeSpecName "kube-api-access-56shc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:28:03 crc kubenswrapper[4894]: I1209 16:28:03.016342 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8448fa91-d4e8-4331-bbe3-a44f64cb1a01-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8448fa91-d4e8-4331-bbe3-a44f64cb1a01" (UID: "8448fa91-d4e8-4331-bbe3-a44f64cb1a01"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:28:03 crc kubenswrapper[4894]: I1209 16:28:03.022873 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-56shc\" (UniqueName: \"kubernetes.io/projected/8448fa91-d4e8-4331-bbe3-a44f64cb1a01-kube-api-access-56shc\") on node \"crc\" DevicePath \"\"" Dec 09 16:28:03 crc kubenswrapper[4894]: I1209 16:28:03.022954 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8448fa91-d4e8-4331-bbe3-a44f64cb1a01-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:28:03 crc kubenswrapper[4894]: I1209 16:28:03.022968 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8448fa91-d4e8-4331-bbe3-a44f64cb1a01-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:28:03 crc kubenswrapper[4894]: I1209 16:28:03.298824 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h4bm2" Dec 09 16:28:03 crc kubenswrapper[4894]: I1209 16:28:03.300829 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h4bm2" event={"ID":"8448fa91-d4e8-4331-bbe3-a44f64cb1a01","Type":"ContainerDied","Data":"11096ffd6574a3263287615bf9499a40ffce890bc18ab568f9994e36dc79a624"} Dec 09 16:28:03 crc kubenswrapper[4894]: I1209 16:28:03.300918 4894 scope.go:117] "RemoveContainer" containerID="ab33013bbace00bb0427d3ac0d23af82ae9561e689b2f75570827dbfd3bf94d1" Dec 09 16:28:03 crc kubenswrapper[4894]: E1209 16:28:03.303382 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tempest-tests-tempest-tests-runner\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified\\\"\"" pod="openstack/tempest-tests-tempest-s00-full" podUID="675a2b9e-a4fe-4ced-9997-c82358f8550c" Dec 09 16:28:03 crc kubenswrapper[4894]: I1209 16:28:03.360251 4894 scope.go:117] "RemoveContainer" containerID="440116e5bf480658d386c6f65f881a19dbd0555ddae038c0771397aedd137173" Dec 09 16:28:03 crc kubenswrapper[4894]: I1209 16:28:03.364445 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h4bm2"] Dec 09 16:28:03 crc kubenswrapper[4894]: I1209 16:28:03.373935 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-h4bm2"] Dec 09 16:28:03 crc kubenswrapper[4894]: I1209 16:28:03.390723 4894 scope.go:117] "RemoveContainer" containerID="d4ce7e8cea22fd5488e76c539b9a879c63073f67de00609b5a48ef718c6b13db" Dec 09 16:28:04 crc kubenswrapper[4894]: I1209 16:28:04.127716 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8448fa91-d4e8-4331-bbe3-a44f64cb1a01" path="/var/lib/kubelet/pods/8448fa91-d4e8-4331-bbe3-a44f64cb1a01/volumes" Dec 09 16:28:12 crc kubenswrapper[4894]: I1209 16:28:12.257957 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:28:12 crc kubenswrapper[4894]: I1209 16:28:12.258697 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:28:19 crc kubenswrapper[4894]: I1209 16:28:19.706788 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s0" Dec 09 16:28:21 crc kubenswrapper[4894]: I1209 16:28:21.519453 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"675a2b9e-a4fe-4ced-9997-c82358f8550c","Type":"ContainerStarted","Data":"1b0d3fca7c66fe155d5720466af95782b0cd85bb40fd9bbb92c30b805ba31838"} Dec 09 16:28:21 crc kubenswrapper[4894]: I1209 16:28:21.550298 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest-s00-full" podStartSLOduration=4.143312627 podStartE2EDuration="54.550268577s" podCreationTimestamp="2025-12-09 16:27:27 +0000 UTC" firstStartedPulling="2025-12-09 16:27:29.250418866 +0000 UTC m=+3343.569629535" lastFinishedPulling="2025-12-09 16:28:19.657374806 +0000 UTC m=+3393.976585485" observedRunningTime="2025-12-09 16:28:21.545227441 +0000 UTC m=+3395.864438120" watchObservedRunningTime="2025-12-09 16:28:21.550268577 +0000 UTC m=+3395.869479276" Dec 09 16:28:42 crc kubenswrapper[4894]: I1209 16:28:42.258029 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:28:42 crc kubenswrapper[4894]: I1209 16:28:42.258529 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:28:57 crc kubenswrapper[4894]: I1209 16:28:57.954272 4894 scope.go:117] "RemoveContainer" containerID="7e118c7416a0c910851a8d1dea860c2c015a132818086b80cc884f76269e4fda" Dec 09 16:28:57 crc kubenswrapper[4894]: I1209 16:28:57.988682 4894 scope.go:117] "RemoveContainer" containerID="6b99ea2b2e4a8ac9a686ab6ae309ef3b88140b90ae86019bca4ba38ad992f6eb" Dec 09 16:28:58 crc kubenswrapper[4894]: I1209 16:28:58.012545 4894 scope.go:117] "RemoveContainer" containerID="2999f4db583377a8063ba25b986928f6b5092f0186c8de910cea657c35695674" Dec 09 16:28:58 crc kubenswrapper[4894]: I1209 16:28:58.029208 4894 scope.go:117] "RemoveContainer" containerID="228da1a824c1215b32b05aa5237aaa5ba91ae22d35f8d03711b8b6d460264daa" Dec 09 16:29:12 crc kubenswrapper[4894]: I1209 16:29:12.258090 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:29:12 crc kubenswrapper[4894]: I1209 16:29:12.258634 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:29:12 crc kubenswrapper[4894]: I1209 16:29:12.258704 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 16:29:12 crc kubenswrapper[4894]: I1209 16:29:12.259528 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"92d4a03025122ef2af3ae2a9840aaf815cef6ed1352687e1bda2ee74c9f07415"} pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 16:29:12 crc kubenswrapper[4894]: I1209 16:29:12.259590 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" containerID="cri-o://92d4a03025122ef2af3ae2a9840aaf815cef6ed1352687e1bda2ee74c9f07415" gracePeriod=600 Dec 09 16:29:13 crc kubenswrapper[4894]: I1209 16:29:13.081940 4894 generic.go:334] "Generic (PLEG): container finished" podID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerID="92d4a03025122ef2af3ae2a9840aaf815cef6ed1352687e1bda2ee74c9f07415" exitCode=0 Dec 09 16:29:13 crc kubenswrapper[4894]: I1209 16:29:13.082084 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerDied","Data":"92d4a03025122ef2af3ae2a9840aaf815cef6ed1352687e1bda2ee74c9f07415"} Dec 09 16:29:13 crc kubenswrapper[4894]: I1209 16:29:13.082605 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerStarted","Data":"5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754"} Dec 09 16:29:13 crc kubenswrapper[4894]: I1209 16:29:13.082631 4894 scope.go:117] "RemoveContainer" containerID="8687a18682fd3b6dcad98faedca09a6e3b80efc99d9e7c543ef3fd17929eef86" Dec 09 16:30:00 crc kubenswrapper[4894]: I1209 16:30:00.177618 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421630-w976r"] Dec 09 16:30:00 crc kubenswrapper[4894]: E1209 16:30:00.178590 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8448fa91-d4e8-4331-bbe3-a44f64cb1a01" containerName="registry-server" Dec 09 16:30:00 crc kubenswrapper[4894]: I1209 16:30:00.178605 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8448fa91-d4e8-4331-bbe3-a44f64cb1a01" containerName="registry-server" Dec 09 16:30:00 crc kubenswrapper[4894]: E1209 16:30:00.178628 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8448fa91-d4e8-4331-bbe3-a44f64cb1a01" containerName="extract-utilities" Dec 09 16:30:00 crc kubenswrapper[4894]: I1209 16:30:00.178655 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8448fa91-d4e8-4331-bbe3-a44f64cb1a01" containerName="extract-utilities" Dec 09 16:30:00 crc kubenswrapper[4894]: E1209 16:30:00.178697 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8448fa91-d4e8-4331-bbe3-a44f64cb1a01" containerName="extract-content" Dec 09 16:30:00 crc kubenswrapper[4894]: I1209 16:30:00.178705 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="8448fa91-d4e8-4331-bbe3-a44f64cb1a01" containerName="extract-content" Dec 09 16:30:00 crc kubenswrapper[4894]: I1209 16:30:00.178916 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="8448fa91-d4e8-4331-bbe3-a44f64cb1a01" containerName="registry-server" Dec 09 16:30:00 crc kubenswrapper[4894]: I1209 16:30:00.179695 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421630-w976r" Dec 09 16:30:00 crc kubenswrapper[4894]: I1209 16:30:00.181729 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 16:30:00 crc kubenswrapper[4894]: I1209 16:30:00.185422 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 16:30:00 crc kubenswrapper[4894]: I1209 16:30:00.186417 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421630-w976r"] Dec 09 16:30:00 crc kubenswrapper[4894]: I1209 16:30:00.227566 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c-config-volume\") pod \"collect-profiles-29421630-w976r\" (UID: \"b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421630-w976r" Dec 09 16:30:00 crc kubenswrapper[4894]: I1209 16:30:00.227743 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c-secret-volume\") pod \"collect-profiles-29421630-w976r\" (UID: \"b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421630-w976r" Dec 09 16:30:00 crc kubenswrapper[4894]: I1209 16:30:00.227829 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q2q97\" (UniqueName: \"kubernetes.io/projected/b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c-kube-api-access-q2q97\") pod \"collect-profiles-29421630-w976r\" (UID: \"b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421630-w976r" Dec 09 16:30:00 crc kubenswrapper[4894]: I1209 16:30:00.328907 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c-config-volume\") pod \"collect-profiles-29421630-w976r\" (UID: \"b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421630-w976r" Dec 09 16:30:00 crc kubenswrapper[4894]: I1209 16:30:00.329010 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c-secret-volume\") pod \"collect-profiles-29421630-w976r\" (UID: \"b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421630-w976r" Dec 09 16:30:00 crc kubenswrapper[4894]: I1209 16:30:00.329063 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q2q97\" (UniqueName: \"kubernetes.io/projected/b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c-kube-api-access-q2q97\") pod \"collect-profiles-29421630-w976r\" (UID: \"b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421630-w976r" Dec 09 16:30:00 crc kubenswrapper[4894]: I1209 16:30:00.330202 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c-config-volume\") pod \"collect-profiles-29421630-w976r\" (UID: \"b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421630-w976r" Dec 09 16:30:00 crc kubenswrapper[4894]: I1209 16:30:00.341126 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c-secret-volume\") pod \"collect-profiles-29421630-w976r\" (UID: \"b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421630-w976r" Dec 09 16:30:00 crc kubenswrapper[4894]: I1209 16:30:00.345062 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q2q97\" (UniqueName: \"kubernetes.io/projected/b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c-kube-api-access-q2q97\") pod \"collect-profiles-29421630-w976r\" (UID: \"b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421630-w976r" Dec 09 16:30:00 crc kubenswrapper[4894]: I1209 16:30:00.506936 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421630-w976r" Dec 09 16:30:01 crc kubenswrapper[4894]: I1209 16:30:01.077987 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421630-w976r"] Dec 09 16:30:01 crc kubenswrapper[4894]: W1209 16:30:01.089521 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb545b4d5_7c03_4e8b_b7fd_cbf1b12fcb7c.slice/crio-b5503a7c176940392b39bc260d9cef69b047d29e5816813eb7e80322a65c2593 WatchSource:0}: Error finding container b5503a7c176940392b39bc260d9cef69b047d29e5816813eb7e80322a65c2593: Status 404 returned error can't find the container with id b5503a7c176940392b39bc260d9cef69b047d29e5816813eb7e80322a65c2593 Dec 09 16:30:01 crc kubenswrapper[4894]: I1209 16:30:01.620889 4894 generic.go:334] "Generic (PLEG): container finished" podID="b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c" containerID="1052c532cffb3a5a8b252bb9805fcb61e158df60c4e0e01cb3c638f2bf0e0b02" exitCode=0 Dec 09 16:30:01 crc kubenswrapper[4894]: I1209 16:30:01.620991 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421630-w976r" event={"ID":"b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c","Type":"ContainerDied","Data":"1052c532cffb3a5a8b252bb9805fcb61e158df60c4e0e01cb3c638f2bf0e0b02"} Dec 09 16:30:01 crc kubenswrapper[4894]: I1209 16:30:01.621242 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421630-w976r" event={"ID":"b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c","Type":"ContainerStarted","Data":"b5503a7c176940392b39bc260d9cef69b047d29e5816813eb7e80322a65c2593"} Dec 09 16:30:03 crc kubenswrapper[4894]: I1209 16:30:03.002785 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421630-w976r" Dec 09 16:30:03 crc kubenswrapper[4894]: I1209 16:30:03.186809 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q2q97\" (UniqueName: \"kubernetes.io/projected/b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c-kube-api-access-q2q97\") pod \"b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c\" (UID: \"b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c\") " Dec 09 16:30:03 crc kubenswrapper[4894]: I1209 16:30:03.186945 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c-config-volume\") pod \"b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c\" (UID: \"b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c\") " Dec 09 16:30:03 crc kubenswrapper[4894]: I1209 16:30:03.186972 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c-secret-volume\") pod \"b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c\" (UID: \"b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c\") " Dec 09 16:30:03 crc kubenswrapper[4894]: I1209 16:30:03.187916 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c-config-volume" (OuterVolumeSpecName: "config-volume") pod "b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c" (UID: "b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:30:03 crc kubenswrapper[4894]: I1209 16:30:03.193285 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c-kube-api-access-q2q97" (OuterVolumeSpecName: "kube-api-access-q2q97") pod "b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c" (UID: "b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c"). InnerVolumeSpecName "kube-api-access-q2q97". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:30:03 crc kubenswrapper[4894]: I1209 16:30:03.202752 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c" (UID: "b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:30:03 crc kubenswrapper[4894]: I1209 16:30:03.289432 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q2q97\" (UniqueName: \"kubernetes.io/projected/b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c-kube-api-access-q2q97\") on node \"crc\" DevicePath \"\"" Dec 09 16:30:03 crc kubenswrapper[4894]: I1209 16:30:03.289478 4894 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 16:30:03 crc kubenswrapper[4894]: I1209 16:30:03.289493 4894 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 16:30:03 crc kubenswrapper[4894]: I1209 16:30:03.641422 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421630-w976r" event={"ID":"b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c","Type":"ContainerDied","Data":"b5503a7c176940392b39bc260d9cef69b047d29e5816813eb7e80322a65c2593"} Dec 09 16:30:03 crc kubenswrapper[4894]: I1209 16:30:03.641695 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b5503a7c176940392b39bc260d9cef69b047d29e5816813eb7e80322a65c2593" Dec 09 16:30:03 crc kubenswrapper[4894]: I1209 16:30:03.641506 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421630-w976r" Dec 09 16:30:04 crc kubenswrapper[4894]: I1209 16:30:04.078751 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421585-wrfw9"] Dec 09 16:30:04 crc kubenswrapper[4894]: I1209 16:30:04.092806 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421585-wrfw9"] Dec 09 16:30:04 crc kubenswrapper[4894]: I1209 16:30:04.117145 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67604b04-833c-4a08-a809-860c3fdbf802" path="/var/lib/kubelet/pods/67604b04-833c-4a08-a809-860c3fdbf802/volumes" Dec 09 16:30:52 crc kubenswrapper[4894]: I1209 16:30:52.154221 4894 generic.go:334] "Generic (PLEG): container finished" podID="675a2b9e-a4fe-4ced-9997-c82358f8550c" containerID="1b0d3fca7c66fe155d5720466af95782b0cd85bb40fd9bbb92c30b805ba31838" exitCode=123 Dec 09 16:30:52 crc kubenswrapper[4894]: I1209 16:30:52.154816 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"675a2b9e-a4fe-4ced-9997-c82358f8550c","Type":"ContainerDied","Data":"1b0d3fca7c66fe155d5720466af95782b0cd85bb40fd9bbb92c30b805ba31838"} Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.293750 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9hb49"] Dec 09 16:30:53 crc kubenswrapper[4894]: E1209 16:30:53.294169 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c" containerName="collect-profiles" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.294181 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c" containerName="collect-profiles" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.294358 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b545b4d5-7c03-4e8b-b7fd-cbf1b12fcb7c" containerName="collect-profiles" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.295939 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9hb49" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.309577 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9hb49"] Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.399084 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9b080e8-a8fd-4595-9f3c-c12106f56501-catalog-content\") pod \"community-operators-9hb49\" (UID: \"c9b080e8-a8fd-4595-9f3c-c12106f56501\") " pod="openshift-marketplace/community-operators-9hb49" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.399309 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55s9x\" (UniqueName: \"kubernetes.io/projected/c9b080e8-a8fd-4595-9f3c-c12106f56501-kube-api-access-55s9x\") pod \"community-operators-9hb49\" (UID: \"c9b080e8-a8fd-4595-9f3c-c12106f56501\") " pod="openshift-marketplace/community-operators-9hb49" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.399360 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9b080e8-a8fd-4595-9f3c-c12106f56501-utilities\") pod \"community-operators-9hb49\" (UID: \"c9b080e8-a8fd-4595-9f3c-c12106f56501\") " pod="openshift-marketplace/community-operators-9hb49" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.501343 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55s9x\" (UniqueName: \"kubernetes.io/projected/c9b080e8-a8fd-4595-9f3c-c12106f56501-kube-api-access-55s9x\") pod \"community-operators-9hb49\" (UID: \"c9b080e8-a8fd-4595-9f3c-c12106f56501\") " pod="openshift-marketplace/community-operators-9hb49" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.501690 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9b080e8-a8fd-4595-9f3c-c12106f56501-utilities\") pod \"community-operators-9hb49\" (UID: \"c9b080e8-a8fd-4595-9f3c-c12106f56501\") " pod="openshift-marketplace/community-operators-9hb49" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.501753 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9b080e8-a8fd-4595-9f3c-c12106f56501-catalog-content\") pod \"community-operators-9hb49\" (UID: \"c9b080e8-a8fd-4595-9f3c-c12106f56501\") " pod="openshift-marketplace/community-operators-9hb49" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.502253 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9b080e8-a8fd-4595-9f3c-c12106f56501-catalog-content\") pod \"community-operators-9hb49\" (UID: \"c9b080e8-a8fd-4595-9f3c-c12106f56501\") " pod="openshift-marketplace/community-operators-9hb49" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.502473 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9b080e8-a8fd-4595-9f3c-c12106f56501-utilities\") pod \"community-operators-9hb49\" (UID: \"c9b080e8-a8fd-4595-9f3c-c12106f56501\") " pod="openshift-marketplace/community-operators-9hb49" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.530951 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55s9x\" (UniqueName: \"kubernetes.io/projected/c9b080e8-a8fd-4595-9f3c-c12106f56501-kube-api-access-55s9x\") pod \"community-operators-9hb49\" (UID: \"c9b080e8-a8fd-4595-9f3c-c12106f56501\") " pod="openshift-marketplace/community-operators-9hb49" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.640158 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9hb49" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.759045 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.848329 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tempest-tests-tempest-s01-single-test"] Dec 09 16:30:53 crc kubenswrapper[4894]: E1209 16:30:53.849154 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="675a2b9e-a4fe-4ced-9997-c82358f8550c" containerName="tempest-tests-tempest-tests-runner" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.849170 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="675a2b9e-a4fe-4ced-9997-c82358f8550c" containerName="tempest-tests-tempest-tests-runner" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.849406 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="675a2b9e-a4fe-4ced-9997-c82358f8550c" containerName="tempest-tests-tempest-tests-runner" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.850511 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.855674 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-custom-data-s1" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.855741 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tempest-tests-tempest-env-vars-s1" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.856033 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s01-single-test"] Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.911681 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/675a2b9e-a4fe-4ced-9997-c82358f8550c-test-operator-ephemeral-workdir\") pod \"675a2b9e-a4fe-4ced-9997-c82358f8550c\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.911739 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-openstack-config-secret\") pod \"675a2b9e-a4fe-4ced-9997-c82358f8550c\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.911763 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/675a2b9e-a4fe-4ced-9997-c82358f8550c-config-data\") pod \"675a2b9e-a4fe-4ced-9997-c82358f8550c\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.911849 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5bm79\" (UniqueName: \"kubernetes.io/projected/675a2b9e-a4fe-4ced-9997-c82358f8550c-kube-api-access-5bm79\") pod \"675a2b9e-a4fe-4ced-9997-c82358f8550c\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.911941 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/675a2b9e-a4fe-4ced-9997-c82358f8550c-test-operator-ephemeral-temporary\") pod \"675a2b9e-a4fe-4ced-9997-c82358f8550c\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.911979 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-ssh-key\") pod \"675a2b9e-a4fe-4ced-9997-c82358f8550c\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.911996 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-ceph\") pod \"675a2b9e-a4fe-4ced-9997-c82358f8550c\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.912051 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-ca-certs\") pod \"675a2b9e-a4fe-4ced-9997-c82358f8550c\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.912076 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"675a2b9e-a4fe-4ced-9997-c82358f8550c\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.912130 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/675a2b9e-a4fe-4ced-9997-c82358f8550c-openstack-config\") pod \"675a2b9e-a4fe-4ced-9997-c82358f8550c\" (UID: \"675a2b9e-a4fe-4ced-9997-c82358f8550c\") " Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.913572 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/675a2b9e-a4fe-4ced-9997-c82358f8550c-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "675a2b9e-a4fe-4ced-9997-c82358f8550c" (UID: "675a2b9e-a4fe-4ced-9997-c82358f8550c"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.914280 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/675a2b9e-a4fe-4ced-9997-c82358f8550c-config-data" (OuterVolumeSpecName: "config-data") pod "675a2b9e-a4fe-4ced-9997-c82358f8550c" (UID: "675a2b9e-a4fe-4ced-9997-c82358f8550c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.926809 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/675a2b9e-a4fe-4ced-9997-c82358f8550c-kube-api-access-5bm79" (OuterVolumeSpecName: "kube-api-access-5bm79") pod "675a2b9e-a4fe-4ced-9997-c82358f8550c" (UID: "675a2b9e-a4fe-4ced-9997-c82358f8550c"). InnerVolumeSpecName "kube-api-access-5bm79". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.929831 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/675a2b9e-a4fe-4ced-9997-c82358f8550c-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "675a2b9e-a4fe-4ced-9997-c82358f8550c" (UID: "675a2b9e-a4fe-4ced-9997-c82358f8550c"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.932263 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "675a2b9e-a4fe-4ced-9997-c82358f8550c" (UID: "675a2b9e-a4fe-4ced-9997-c82358f8550c"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.932362 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-ceph" (OuterVolumeSpecName: "ceph") pod "675a2b9e-a4fe-4ced-9997-c82358f8550c" (UID: "675a2b9e-a4fe-4ced-9997-c82358f8550c"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.950605 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "675a2b9e-a4fe-4ced-9997-c82358f8550c" (UID: "675a2b9e-a4fe-4ced-9997-c82358f8550c"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.951206 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "675a2b9e-a4fe-4ced-9997-c82358f8550c" (UID: "675a2b9e-a4fe-4ced-9997-c82358f8550c"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.962129 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "675a2b9e-a4fe-4ced-9997-c82358f8550c" (UID: "675a2b9e-a4fe-4ced-9997-c82358f8550c"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:30:53 crc kubenswrapper[4894]: I1209 16:30:53.975305 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/675a2b9e-a4fe-4ced-9997-c82358f8550c-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "675a2b9e-a4fe-4ced-9997-c82358f8550c" (UID: "675a2b9e-a4fe-4ced-9997-c82358f8550c"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.009710 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9hb49"] Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.014062 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4a49e9ae-60a1-4165-ae4d-c877537d8825-config-data\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.014126 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.014892 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4a49e9ae-60a1-4165-ae4d-c877537d8825-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.014975 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4a49e9ae-60a1-4165-ae4d-c877537d8825-openstack-config\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.015024 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4a49e9ae-60a1-4165-ae4d-c877537d8825-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.015158 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.015199 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-ssh-key\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.015272 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-ceph\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.015341 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-ca-certs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.015377 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xvlqh\" (UniqueName: \"kubernetes.io/projected/4a49e9ae-60a1-4165-ae4d-c877537d8825-kube-api-access-xvlqh\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.015472 4894 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.015498 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/675a2b9e-a4fe-4ced-9997-c82358f8550c-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.015515 4894 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/675a2b9e-a4fe-4ced-9997-c82358f8550c-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.015527 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.015539 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/675a2b9e-a4fe-4ced-9997-c82358f8550c-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.015552 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5bm79\" (UniqueName: \"kubernetes.io/projected/675a2b9e-a4fe-4ced-9997-c82358f8550c-kube-api-access-5bm79\") on node \"crc\" DevicePath \"\"" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.015567 4894 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/675a2b9e-a4fe-4ced-9997-c82358f8550c-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.015769 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.015793 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/675a2b9e-a4fe-4ced-9997-c82358f8550c-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.043488 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.118017 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4a49e9ae-60a1-4165-ae4d-c877537d8825-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.118067 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4a49e9ae-60a1-4165-ae4d-c877537d8825-openstack-config\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.118091 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4a49e9ae-60a1-4165-ae4d-c877537d8825-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.118144 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.118164 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-ssh-key\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.118193 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-ceph\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.118225 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-ca-certs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.118248 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xvlqh\" (UniqueName: \"kubernetes.io/projected/4a49e9ae-60a1-4165-ae4d-c877537d8825-kube-api-access-xvlqh\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.118275 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4a49e9ae-60a1-4165-ae4d-c877537d8825-config-data\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.119112 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4a49e9ae-60a1-4165-ae4d-c877537d8825-test-operator-ephemeral-temporary\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.119619 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4a49e9ae-60a1-4165-ae4d-c877537d8825-config-data\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.120129 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4a49e9ae-60a1-4165-ae4d-c877537d8825-test-operator-ephemeral-workdir\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.120867 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4a49e9ae-60a1-4165-ae4d-c877537d8825-openstack-config\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.124009 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-ssh-key\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.124923 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-ceph\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.125053 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-ca-certs\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.126171 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-openstack-config-secret\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.135327 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xvlqh\" (UniqueName: \"kubernetes.io/projected/4a49e9ae-60a1-4165-ae4d-c877537d8825-kube-api-access-xvlqh\") pod \"tempest-tests-tempest-s01-single-test\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.174258 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hb49" event={"ID":"c9b080e8-a8fd-4595-9f3c-c12106f56501","Type":"ContainerStarted","Data":"ec44fa347d4759f1bde7bd1d2f596534a770fdc38da8b892b099f658ef1f5898"} Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.174577 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hb49" event={"ID":"c9b080e8-a8fd-4595-9f3c-c12106f56501","Type":"ContainerStarted","Data":"037f3c82dfcbabfe6fa52299fa7d3ca3ef69db942bfd10b3ebc89028218b8290"} Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.174364 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.179513 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s00-full" event={"ID":"675a2b9e-a4fe-4ced-9997-c82358f8550c","Type":"ContainerDied","Data":"6df7e934fdfad472b46e55e1e2592ac3e1dd73434babb8fb1d78c184226db643"} Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.179551 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6df7e934fdfad472b46e55e1e2592ac3e1dd73434babb8fb1d78c184226db643" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.179623 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s00-full" Dec 09 16:30:54 crc kubenswrapper[4894]: I1209 16:30:54.707194 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tempest-tests-tempest-s01-single-test"] Dec 09 16:30:54 crc kubenswrapper[4894]: W1209 16:30:54.712155 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4a49e9ae_60a1_4165_ae4d_c877537d8825.slice/crio-16d24c88b9e0f80169873c3c2bdbada7e879182674ed73d50a17afb842091c4b WatchSource:0}: Error finding container 16d24c88b9e0f80169873c3c2bdbada7e879182674ed73d50a17afb842091c4b: Status 404 returned error can't find the container with id 16d24c88b9e0f80169873c3c2bdbada7e879182674ed73d50a17afb842091c4b Dec 09 16:30:55 crc kubenswrapper[4894]: I1209 16:30:55.188500 4894 generic.go:334] "Generic (PLEG): container finished" podID="c9b080e8-a8fd-4595-9f3c-c12106f56501" containerID="ec44fa347d4759f1bde7bd1d2f596534a770fdc38da8b892b099f658ef1f5898" exitCode=0 Dec 09 16:30:55 crc kubenswrapper[4894]: I1209 16:30:55.188569 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hb49" event={"ID":"c9b080e8-a8fd-4595-9f3c-c12106f56501","Type":"ContainerDied","Data":"ec44fa347d4759f1bde7bd1d2f596534a770fdc38da8b892b099f658ef1f5898"} Dec 09 16:30:55 crc kubenswrapper[4894]: I1209 16:30:55.190307 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 16:30:55 crc kubenswrapper[4894]: I1209 16:30:55.191228 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"4a49e9ae-60a1-4165-ae4d-c877537d8825","Type":"ContainerStarted","Data":"16d24c88b9e0f80169873c3c2bdbada7e879182674ed73d50a17afb842091c4b"} Dec 09 16:30:56 crc kubenswrapper[4894]: I1209 16:30:56.221804 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"4a49e9ae-60a1-4165-ae4d-c877537d8825","Type":"ContainerStarted","Data":"de3d7b6989d294b0e7fac090792ffd6f5e333ed38a4800fdfce3729860b2f126"} Dec 09 16:30:56 crc kubenswrapper[4894]: I1209 16:30:56.248846 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tempest-tests-tempest-s01-single-test" podStartSLOduration=3.248823523 podStartE2EDuration="3.248823523s" podCreationTimestamp="2025-12-09 16:30:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:30:56.248098023 +0000 UTC m=+3550.567308702" watchObservedRunningTime="2025-12-09 16:30:56.248823523 +0000 UTC m=+3550.568034192" Dec 09 16:30:57 crc kubenswrapper[4894]: I1209 16:30:57.234343 4894 generic.go:334] "Generic (PLEG): container finished" podID="c9b080e8-a8fd-4595-9f3c-c12106f56501" containerID="628fe7672399145433b04da57374d329b40e612a3addf0daa75f74f0c8f83bd7" exitCode=0 Dec 09 16:30:57 crc kubenswrapper[4894]: I1209 16:30:57.234422 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hb49" event={"ID":"c9b080e8-a8fd-4595-9f3c-c12106f56501","Type":"ContainerDied","Data":"628fe7672399145433b04da57374d329b40e612a3addf0daa75f74f0c8f83bd7"} Dec 09 16:30:58 crc kubenswrapper[4894]: I1209 16:30:58.133520 4894 scope.go:117] "RemoveContainer" containerID="03d5d47017f84d6610e92b5ac3f39af962e1fb0802f4fe13bd6b393b4d0aeff7" Dec 09 16:30:58 crc kubenswrapper[4894]: I1209 16:30:58.243892 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hb49" event={"ID":"c9b080e8-a8fd-4595-9f3c-c12106f56501","Type":"ContainerStarted","Data":"53f08e55cbaeb9a39a53053bacb61384ff866e16fe373d8b15c8ad3971c1e716"} Dec 09 16:30:58 crc kubenswrapper[4894]: I1209 16:30:58.266772 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9hb49" podStartSLOduration=2.7111188090000002 podStartE2EDuration="5.266754589s" podCreationTimestamp="2025-12-09 16:30:53 +0000 UTC" firstStartedPulling="2025-12-09 16:30:55.190051209 +0000 UTC m=+3549.509261878" lastFinishedPulling="2025-12-09 16:30:57.745686989 +0000 UTC m=+3552.064897658" observedRunningTime="2025-12-09 16:30:58.261248259 +0000 UTC m=+3552.580458918" watchObservedRunningTime="2025-12-09 16:30:58.266754589 +0000 UTC m=+3552.585965258" Dec 09 16:31:03 crc kubenswrapper[4894]: I1209 16:31:03.640746 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9hb49" Dec 09 16:31:03 crc kubenswrapper[4894]: I1209 16:31:03.641417 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9hb49" Dec 09 16:31:03 crc kubenswrapper[4894]: I1209 16:31:03.709729 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9hb49" Dec 09 16:31:04 crc kubenswrapper[4894]: I1209 16:31:04.371290 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9hb49" Dec 09 16:31:04 crc kubenswrapper[4894]: I1209 16:31:04.434814 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9hb49"] Dec 09 16:31:06 crc kubenswrapper[4894]: I1209 16:31:06.335202 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9hb49" podUID="c9b080e8-a8fd-4595-9f3c-c12106f56501" containerName="registry-server" containerID="cri-o://53f08e55cbaeb9a39a53053bacb61384ff866e16fe373d8b15c8ad3971c1e716" gracePeriod=2 Dec 09 16:31:06 crc kubenswrapper[4894]: I1209 16:31:06.992535 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9hb49" Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.094416 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9b080e8-a8fd-4595-9f3c-c12106f56501-utilities\") pod \"c9b080e8-a8fd-4595-9f3c-c12106f56501\" (UID: \"c9b080e8-a8fd-4595-9f3c-c12106f56501\") " Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.094521 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55s9x\" (UniqueName: \"kubernetes.io/projected/c9b080e8-a8fd-4595-9f3c-c12106f56501-kube-api-access-55s9x\") pod \"c9b080e8-a8fd-4595-9f3c-c12106f56501\" (UID: \"c9b080e8-a8fd-4595-9f3c-c12106f56501\") " Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.094758 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9b080e8-a8fd-4595-9f3c-c12106f56501-catalog-content\") pod \"c9b080e8-a8fd-4595-9f3c-c12106f56501\" (UID: \"c9b080e8-a8fd-4595-9f3c-c12106f56501\") " Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.095854 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9b080e8-a8fd-4595-9f3c-c12106f56501-utilities" (OuterVolumeSpecName: "utilities") pod "c9b080e8-a8fd-4595-9f3c-c12106f56501" (UID: "c9b080e8-a8fd-4595-9f3c-c12106f56501"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.096360 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c9b080e8-a8fd-4595-9f3c-c12106f56501-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.100950 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c9b080e8-a8fd-4595-9f3c-c12106f56501-kube-api-access-55s9x" (OuterVolumeSpecName: "kube-api-access-55s9x") pod "c9b080e8-a8fd-4595-9f3c-c12106f56501" (UID: "c9b080e8-a8fd-4595-9f3c-c12106f56501"). InnerVolumeSpecName "kube-api-access-55s9x". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.151321 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c9b080e8-a8fd-4595-9f3c-c12106f56501-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c9b080e8-a8fd-4595-9f3c-c12106f56501" (UID: "c9b080e8-a8fd-4595-9f3c-c12106f56501"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.199380 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c9b080e8-a8fd-4595-9f3c-c12106f56501-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.199408 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55s9x\" (UniqueName: \"kubernetes.io/projected/c9b080e8-a8fd-4595-9f3c-c12106f56501-kube-api-access-55s9x\") on node \"crc\" DevicePath \"\"" Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.350292 4894 generic.go:334] "Generic (PLEG): container finished" podID="c9b080e8-a8fd-4595-9f3c-c12106f56501" containerID="53f08e55cbaeb9a39a53053bacb61384ff866e16fe373d8b15c8ad3971c1e716" exitCode=0 Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.350361 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hb49" event={"ID":"c9b080e8-a8fd-4595-9f3c-c12106f56501","Type":"ContainerDied","Data":"53f08e55cbaeb9a39a53053bacb61384ff866e16fe373d8b15c8ad3971c1e716"} Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.350422 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9hb49" Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.350745 4894 scope.go:117] "RemoveContainer" containerID="53f08e55cbaeb9a39a53053bacb61384ff866e16fe373d8b15c8ad3971c1e716" Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.350724 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9hb49" event={"ID":"c9b080e8-a8fd-4595-9f3c-c12106f56501","Type":"ContainerDied","Data":"037f3c82dfcbabfe6fa52299fa7d3ca3ef69db942bfd10b3ebc89028218b8290"} Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.410010 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9hb49"] Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.416394 4894 scope.go:117] "RemoveContainer" containerID="628fe7672399145433b04da57374d329b40e612a3addf0daa75f74f0c8f83bd7" Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.419813 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9hb49"] Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.445945 4894 scope.go:117] "RemoveContainer" containerID="ec44fa347d4759f1bde7bd1d2f596534a770fdc38da8b892b099f658ef1f5898" Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.484233 4894 scope.go:117] "RemoveContainer" containerID="53f08e55cbaeb9a39a53053bacb61384ff866e16fe373d8b15c8ad3971c1e716" Dec 09 16:31:07 crc kubenswrapper[4894]: E1209 16:31:07.485988 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53f08e55cbaeb9a39a53053bacb61384ff866e16fe373d8b15c8ad3971c1e716\": container with ID starting with 53f08e55cbaeb9a39a53053bacb61384ff866e16fe373d8b15c8ad3971c1e716 not found: ID does not exist" containerID="53f08e55cbaeb9a39a53053bacb61384ff866e16fe373d8b15c8ad3971c1e716" Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.486038 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53f08e55cbaeb9a39a53053bacb61384ff866e16fe373d8b15c8ad3971c1e716"} err="failed to get container status \"53f08e55cbaeb9a39a53053bacb61384ff866e16fe373d8b15c8ad3971c1e716\": rpc error: code = NotFound desc = could not find container \"53f08e55cbaeb9a39a53053bacb61384ff866e16fe373d8b15c8ad3971c1e716\": container with ID starting with 53f08e55cbaeb9a39a53053bacb61384ff866e16fe373d8b15c8ad3971c1e716 not found: ID does not exist" Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.486068 4894 scope.go:117] "RemoveContainer" containerID="628fe7672399145433b04da57374d329b40e612a3addf0daa75f74f0c8f83bd7" Dec 09 16:31:07 crc kubenswrapper[4894]: E1209 16:31:07.486567 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"628fe7672399145433b04da57374d329b40e612a3addf0daa75f74f0c8f83bd7\": container with ID starting with 628fe7672399145433b04da57374d329b40e612a3addf0daa75f74f0c8f83bd7 not found: ID does not exist" containerID="628fe7672399145433b04da57374d329b40e612a3addf0daa75f74f0c8f83bd7" Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.486616 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"628fe7672399145433b04da57374d329b40e612a3addf0daa75f74f0c8f83bd7"} err="failed to get container status \"628fe7672399145433b04da57374d329b40e612a3addf0daa75f74f0c8f83bd7\": rpc error: code = NotFound desc = could not find container \"628fe7672399145433b04da57374d329b40e612a3addf0daa75f74f0c8f83bd7\": container with ID starting with 628fe7672399145433b04da57374d329b40e612a3addf0daa75f74f0c8f83bd7 not found: ID does not exist" Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.486672 4894 scope.go:117] "RemoveContainer" containerID="ec44fa347d4759f1bde7bd1d2f596534a770fdc38da8b892b099f658ef1f5898" Dec 09 16:31:07 crc kubenswrapper[4894]: E1209 16:31:07.487036 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec44fa347d4759f1bde7bd1d2f596534a770fdc38da8b892b099f658ef1f5898\": container with ID starting with ec44fa347d4759f1bde7bd1d2f596534a770fdc38da8b892b099f658ef1f5898 not found: ID does not exist" containerID="ec44fa347d4759f1bde7bd1d2f596534a770fdc38da8b892b099f658ef1f5898" Dec 09 16:31:07 crc kubenswrapper[4894]: I1209 16:31:07.487077 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec44fa347d4759f1bde7bd1d2f596534a770fdc38da8b892b099f658ef1f5898"} err="failed to get container status \"ec44fa347d4759f1bde7bd1d2f596534a770fdc38da8b892b099f658ef1f5898\": rpc error: code = NotFound desc = could not find container \"ec44fa347d4759f1bde7bd1d2f596534a770fdc38da8b892b099f658ef1f5898\": container with ID starting with ec44fa347d4759f1bde7bd1d2f596534a770fdc38da8b892b099f658ef1f5898 not found: ID does not exist" Dec 09 16:31:08 crc kubenswrapper[4894]: I1209 16:31:08.124389 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c9b080e8-a8fd-4595-9f3c-c12106f56501" path="/var/lib/kubelet/pods/c9b080e8-a8fd-4595-9f3c-c12106f56501/volumes" Dec 09 16:31:12 crc kubenswrapper[4894]: I1209 16:31:12.258081 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:31:12 crc kubenswrapper[4894]: I1209 16:31:12.258515 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:31:42 crc kubenswrapper[4894]: I1209 16:31:42.257698 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:31:42 crc kubenswrapper[4894]: I1209 16:31:42.258492 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:32:12 crc kubenswrapper[4894]: I1209 16:32:12.257506 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:32:12 crc kubenswrapper[4894]: I1209 16:32:12.258039 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:32:12 crc kubenswrapper[4894]: I1209 16:32:12.258087 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 16:32:12 crc kubenswrapper[4894]: I1209 16:32:12.258630 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754"} pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 16:32:12 crc kubenswrapper[4894]: I1209 16:32:12.258699 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" containerID="cri-o://5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" gracePeriod=600 Dec 09 16:32:12 crc kubenswrapper[4894]: E1209 16:32:12.390878 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:32:13 crc kubenswrapper[4894]: I1209 16:32:13.047238 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-bf42-account-create-update-6qjdg"] Dec 09 16:32:13 crc kubenswrapper[4894]: I1209 16:32:13.058036 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-create-rn6kt"] Dec 09 16:32:13 crc kubenswrapper[4894]: I1209 16:32:13.069518 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-bf42-account-create-update-6qjdg"] Dec 09 16:32:13 crc kubenswrapper[4894]: I1209 16:32:13.072692 4894 generic.go:334] "Generic (PLEG): container finished" podID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" exitCode=0 Dec 09 16:32:13 crc kubenswrapper[4894]: I1209 16:32:13.072743 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerDied","Data":"5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754"} Dec 09 16:32:13 crc kubenswrapper[4894]: I1209 16:32:13.072811 4894 scope.go:117] "RemoveContainer" containerID="92d4a03025122ef2af3ae2a9840aaf815cef6ed1352687e1bda2ee74c9f07415" Dec 09 16:32:13 crc kubenswrapper[4894]: I1209 16:32:13.073423 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:32:13 crc kubenswrapper[4894]: E1209 16:32:13.073741 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:32:13 crc kubenswrapper[4894]: I1209 16:32:13.079507 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-create-rn6kt"] Dec 09 16:32:14 crc kubenswrapper[4894]: I1209 16:32:14.121143 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a3e1b2cf-f96b-450b-86be-f935ae75d0f0" path="/var/lib/kubelet/pods/a3e1b2cf-f96b-450b-86be-f935ae75d0f0/volumes" Dec 09 16:32:14 crc kubenswrapper[4894]: I1209 16:32:14.122329 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa0ee67f-8fa7-42e0-a524-0bf863f1bc67" path="/var/lib/kubelet/pods/fa0ee67f-8fa7-42e0-a524-0bf863f1bc67/volumes" Dec 09 16:32:27 crc kubenswrapper[4894]: I1209 16:32:27.107824 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:32:27 crc kubenswrapper[4894]: E1209 16:32:27.108618 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:32:35 crc kubenswrapper[4894]: I1209 16:32:35.060476 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/manila-db-sync-jvrwg"] Dec 09 16:32:35 crc kubenswrapper[4894]: I1209 16:32:35.073506 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/manila-db-sync-jvrwg"] Dec 09 16:32:36 crc kubenswrapper[4894]: I1209 16:32:36.131538 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f88f6fb-7da5-47f6-b0b5-a33af1b6230a" path="/var/lib/kubelet/pods/1f88f6fb-7da5-47f6-b0b5-a33af1b6230a/volumes" Dec 09 16:32:36 crc kubenswrapper[4894]: I1209 16:32:36.844309 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-25ffg"] Dec 09 16:32:36 crc kubenswrapper[4894]: E1209 16:32:36.845118 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9b080e8-a8fd-4595-9f3c-c12106f56501" containerName="extract-utilities" Dec 09 16:32:36 crc kubenswrapper[4894]: I1209 16:32:36.845142 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9b080e8-a8fd-4595-9f3c-c12106f56501" containerName="extract-utilities" Dec 09 16:32:36 crc kubenswrapper[4894]: E1209 16:32:36.845184 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9b080e8-a8fd-4595-9f3c-c12106f56501" containerName="registry-server" Dec 09 16:32:36 crc kubenswrapper[4894]: I1209 16:32:36.845192 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9b080e8-a8fd-4595-9f3c-c12106f56501" containerName="registry-server" Dec 09 16:32:36 crc kubenswrapper[4894]: E1209 16:32:36.845215 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c9b080e8-a8fd-4595-9f3c-c12106f56501" containerName="extract-content" Dec 09 16:32:36 crc kubenswrapper[4894]: I1209 16:32:36.845222 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c9b080e8-a8fd-4595-9f3c-c12106f56501" containerName="extract-content" Dec 09 16:32:36 crc kubenswrapper[4894]: I1209 16:32:36.845427 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="c9b080e8-a8fd-4595-9f3c-c12106f56501" containerName="registry-server" Dec 09 16:32:36 crc kubenswrapper[4894]: I1209 16:32:36.846861 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25ffg" Dec 09 16:32:36 crc kubenswrapper[4894]: I1209 16:32:36.870438 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-25ffg"] Dec 09 16:32:36 crc kubenswrapper[4894]: I1209 16:32:36.947260 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76ae52d7-b36a-4967-9c30-6c3fc5c9e532-catalog-content\") pod \"redhat-marketplace-25ffg\" (UID: \"76ae52d7-b36a-4967-9c30-6c3fc5c9e532\") " pod="openshift-marketplace/redhat-marketplace-25ffg" Dec 09 16:32:36 crc kubenswrapper[4894]: I1209 16:32:36.947341 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qnvk\" (UniqueName: \"kubernetes.io/projected/76ae52d7-b36a-4967-9c30-6c3fc5c9e532-kube-api-access-9qnvk\") pod \"redhat-marketplace-25ffg\" (UID: \"76ae52d7-b36a-4967-9c30-6c3fc5c9e532\") " pod="openshift-marketplace/redhat-marketplace-25ffg" Dec 09 16:32:36 crc kubenswrapper[4894]: I1209 16:32:36.947526 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76ae52d7-b36a-4967-9c30-6c3fc5c9e532-utilities\") pod \"redhat-marketplace-25ffg\" (UID: \"76ae52d7-b36a-4967-9c30-6c3fc5c9e532\") " pod="openshift-marketplace/redhat-marketplace-25ffg" Dec 09 16:32:37 crc kubenswrapper[4894]: I1209 16:32:37.049322 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76ae52d7-b36a-4967-9c30-6c3fc5c9e532-utilities\") pod \"redhat-marketplace-25ffg\" (UID: \"76ae52d7-b36a-4967-9c30-6c3fc5c9e532\") " pod="openshift-marketplace/redhat-marketplace-25ffg" Dec 09 16:32:37 crc kubenswrapper[4894]: I1209 16:32:37.049485 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76ae52d7-b36a-4967-9c30-6c3fc5c9e532-catalog-content\") pod \"redhat-marketplace-25ffg\" (UID: \"76ae52d7-b36a-4967-9c30-6c3fc5c9e532\") " pod="openshift-marketplace/redhat-marketplace-25ffg" Dec 09 16:32:37 crc kubenswrapper[4894]: I1209 16:32:37.049547 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qnvk\" (UniqueName: \"kubernetes.io/projected/76ae52d7-b36a-4967-9c30-6c3fc5c9e532-kube-api-access-9qnvk\") pod \"redhat-marketplace-25ffg\" (UID: \"76ae52d7-b36a-4967-9c30-6c3fc5c9e532\") " pod="openshift-marketplace/redhat-marketplace-25ffg" Dec 09 16:32:37 crc kubenswrapper[4894]: I1209 16:32:37.049992 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76ae52d7-b36a-4967-9c30-6c3fc5c9e532-utilities\") pod \"redhat-marketplace-25ffg\" (UID: \"76ae52d7-b36a-4967-9c30-6c3fc5c9e532\") " pod="openshift-marketplace/redhat-marketplace-25ffg" Dec 09 16:32:37 crc kubenswrapper[4894]: I1209 16:32:37.050076 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76ae52d7-b36a-4967-9c30-6c3fc5c9e532-catalog-content\") pod \"redhat-marketplace-25ffg\" (UID: \"76ae52d7-b36a-4967-9c30-6c3fc5c9e532\") " pod="openshift-marketplace/redhat-marketplace-25ffg" Dec 09 16:32:37 crc kubenswrapper[4894]: I1209 16:32:37.080229 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qnvk\" (UniqueName: \"kubernetes.io/projected/76ae52d7-b36a-4967-9c30-6c3fc5c9e532-kube-api-access-9qnvk\") pod \"redhat-marketplace-25ffg\" (UID: \"76ae52d7-b36a-4967-9c30-6c3fc5c9e532\") " pod="openshift-marketplace/redhat-marketplace-25ffg" Dec 09 16:32:37 crc kubenswrapper[4894]: I1209 16:32:37.170873 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25ffg" Dec 09 16:32:37 crc kubenswrapper[4894]: I1209 16:32:37.682054 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-25ffg"] Dec 09 16:32:38 crc kubenswrapper[4894]: I1209 16:32:38.356055 4894 generic.go:334] "Generic (PLEG): container finished" podID="76ae52d7-b36a-4967-9c30-6c3fc5c9e532" containerID="6c59556d88bde064c2bd4442639bae5d73c2a1a77f784aacf885c144dab00646" exitCode=0 Dec 09 16:32:38 crc kubenswrapper[4894]: I1209 16:32:38.356190 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25ffg" event={"ID":"76ae52d7-b36a-4967-9c30-6c3fc5c9e532","Type":"ContainerDied","Data":"6c59556d88bde064c2bd4442639bae5d73c2a1a77f784aacf885c144dab00646"} Dec 09 16:32:38 crc kubenswrapper[4894]: I1209 16:32:38.356456 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25ffg" event={"ID":"76ae52d7-b36a-4967-9c30-6c3fc5c9e532","Type":"ContainerStarted","Data":"ebf27609601b93f2b9f7036dea0447a5c8f77230eecdab7babd0a8b90f6c804e"} Dec 09 16:32:40 crc kubenswrapper[4894]: I1209 16:32:40.383477 4894 generic.go:334] "Generic (PLEG): container finished" podID="76ae52d7-b36a-4967-9c30-6c3fc5c9e532" containerID="34ec16ac1369c51f3afef3867ecdd197a5373ec6c97571c061a898e7735cc86d" exitCode=0 Dec 09 16:32:40 crc kubenswrapper[4894]: I1209 16:32:40.383619 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25ffg" event={"ID":"76ae52d7-b36a-4967-9c30-6c3fc5c9e532","Type":"ContainerDied","Data":"34ec16ac1369c51f3afef3867ecdd197a5373ec6c97571c061a898e7735cc86d"} Dec 09 16:32:41 crc kubenswrapper[4894]: I1209 16:32:41.397142 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25ffg" event={"ID":"76ae52d7-b36a-4967-9c30-6c3fc5c9e532","Type":"ContainerStarted","Data":"6cbf7769cee6f4aceb7a7230aad5c9533e6e4b6b17a96ad99d31879537a96a50"} Dec 09 16:32:41 crc kubenswrapper[4894]: I1209 16:32:41.429707 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-25ffg" podStartSLOduration=2.866765591 podStartE2EDuration="5.429686748s" podCreationTimestamp="2025-12-09 16:32:36 +0000 UTC" firstStartedPulling="2025-12-09 16:32:38.359919498 +0000 UTC m=+3652.679130207" lastFinishedPulling="2025-12-09 16:32:40.922840685 +0000 UTC m=+3655.242051364" observedRunningTime="2025-12-09 16:32:41.421943407 +0000 UTC m=+3655.741154086" watchObservedRunningTime="2025-12-09 16:32:41.429686748 +0000 UTC m=+3655.748897417" Dec 09 16:32:42 crc kubenswrapper[4894]: I1209 16:32:42.108032 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:32:42 crc kubenswrapper[4894]: E1209 16:32:42.108324 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:32:47 crc kubenswrapper[4894]: I1209 16:32:47.172220 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-25ffg" Dec 09 16:32:47 crc kubenswrapper[4894]: I1209 16:32:47.172756 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-25ffg" Dec 09 16:32:47 crc kubenswrapper[4894]: I1209 16:32:47.212850 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-25ffg" Dec 09 16:32:47 crc kubenswrapper[4894]: I1209 16:32:47.534272 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-25ffg" Dec 09 16:32:47 crc kubenswrapper[4894]: I1209 16:32:47.584618 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-25ffg"] Dec 09 16:32:49 crc kubenswrapper[4894]: I1209 16:32:49.475461 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-25ffg" podUID="76ae52d7-b36a-4967-9c30-6c3fc5c9e532" containerName="registry-server" containerID="cri-o://6cbf7769cee6f4aceb7a7230aad5c9533e6e4b6b17a96ad99d31879537a96a50" gracePeriod=2 Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.273883 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25ffg" Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.436505 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76ae52d7-b36a-4967-9c30-6c3fc5c9e532-catalog-content\") pod \"76ae52d7-b36a-4967-9c30-6c3fc5c9e532\" (UID: \"76ae52d7-b36a-4967-9c30-6c3fc5c9e532\") " Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.436770 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76ae52d7-b36a-4967-9c30-6c3fc5c9e532-utilities\") pod \"76ae52d7-b36a-4967-9c30-6c3fc5c9e532\" (UID: \"76ae52d7-b36a-4967-9c30-6c3fc5c9e532\") " Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.436819 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9qnvk\" (UniqueName: \"kubernetes.io/projected/76ae52d7-b36a-4967-9c30-6c3fc5c9e532-kube-api-access-9qnvk\") pod \"76ae52d7-b36a-4967-9c30-6c3fc5c9e532\" (UID: \"76ae52d7-b36a-4967-9c30-6c3fc5c9e532\") " Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.438889 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76ae52d7-b36a-4967-9c30-6c3fc5c9e532-utilities" (OuterVolumeSpecName: "utilities") pod "76ae52d7-b36a-4967-9c30-6c3fc5c9e532" (UID: "76ae52d7-b36a-4967-9c30-6c3fc5c9e532"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.446823 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/76ae52d7-b36a-4967-9c30-6c3fc5c9e532-kube-api-access-9qnvk" (OuterVolumeSpecName: "kube-api-access-9qnvk") pod "76ae52d7-b36a-4967-9c30-6c3fc5c9e532" (UID: "76ae52d7-b36a-4967-9c30-6c3fc5c9e532"). InnerVolumeSpecName "kube-api-access-9qnvk". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.458045 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/76ae52d7-b36a-4967-9c30-6c3fc5c9e532-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "76ae52d7-b36a-4967-9c30-6c3fc5c9e532" (UID: "76ae52d7-b36a-4967-9c30-6c3fc5c9e532"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.484817 4894 generic.go:334] "Generic (PLEG): container finished" podID="76ae52d7-b36a-4967-9c30-6c3fc5c9e532" containerID="6cbf7769cee6f4aceb7a7230aad5c9533e6e4b6b17a96ad99d31879537a96a50" exitCode=0 Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.484860 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25ffg" event={"ID":"76ae52d7-b36a-4967-9c30-6c3fc5c9e532","Type":"ContainerDied","Data":"6cbf7769cee6f4aceb7a7230aad5c9533e6e4b6b17a96ad99d31879537a96a50"} Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.484882 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-25ffg" Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.484900 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-25ffg" event={"ID":"76ae52d7-b36a-4967-9c30-6c3fc5c9e532","Type":"ContainerDied","Data":"ebf27609601b93f2b9f7036dea0447a5c8f77230eecdab7babd0a8b90f6c804e"} Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.484924 4894 scope.go:117] "RemoveContainer" containerID="6cbf7769cee6f4aceb7a7230aad5c9533e6e4b6b17a96ad99d31879537a96a50" Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.511122 4894 scope.go:117] "RemoveContainer" containerID="34ec16ac1369c51f3afef3867ecdd197a5373ec6c97571c061a898e7735cc86d" Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.530859 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-25ffg"] Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.539847 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/76ae52d7-b36a-4967-9c30-6c3fc5c9e532-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.539872 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/76ae52d7-b36a-4967-9c30-6c3fc5c9e532-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.539883 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9qnvk\" (UniqueName: \"kubernetes.io/projected/76ae52d7-b36a-4967-9c30-6c3fc5c9e532-kube-api-access-9qnvk\") on node \"crc\" DevicePath \"\"" Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.541166 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-25ffg"] Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.552377 4894 scope.go:117] "RemoveContainer" containerID="6c59556d88bde064c2bd4442639bae5d73c2a1a77f784aacf885c144dab00646" Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.582047 4894 scope.go:117] "RemoveContainer" containerID="6cbf7769cee6f4aceb7a7230aad5c9533e6e4b6b17a96ad99d31879537a96a50" Dec 09 16:32:50 crc kubenswrapper[4894]: E1209 16:32:50.582917 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cbf7769cee6f4aceb7a7230aad5c9533e6e4b6b17a96ad99d31879537a96a50\": container with ID starting with 6cbf7769cee6f4aceb7a7230aad5c9533e6e4b6b17a96ad99d31879537a96a50 not found: ID does not exist" containerID="6cbf7769cee6f4aceb7a7230aad5c9533e6e4b6b17a96ad99d31879537a96a50" Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.582967 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cbf7769cee6f4aceb7a7230aad5c9533e6e4b6b17a96ad99d31879537a96a50"} err="failed to get container status \"6cbf7769cee6f4aceb7a7230aad5c9533e6e4b6b17a96ad99d31879537a96a50\": rpc error: code = NotFound desc = could not find container \"6cbf7769cee6f4aceb7a7230aad5c9533e6e4b6b17a96ad99d31879537a96a50\": container with ID starting with 6cbf7769cee6f4aceb7a7230aad5c9533e6e4b6b17a96ad99d31879537a96a50 not found: ID does not exist" Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.582994 4894 scope.go:117] "RemoveContainer" containerID="34ec16ac1369c51f3afef3867ecdd197a5373ec6c97571c061a898e7735cc86d" Dec 09 16:32:50 crc kubenswrapper[4894]: E1209 16:32:50.583436 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"34ec16ac1369c51f3afef3867ecdd197a5373ec6c97571c061a898e7735cc86d\": container with ID starting with 34ec16ac1369c51f3afef3867ecdd197a5373ec6c97571c061a898e7735cc86d not found: ID does not exist" containerID="34ec16ac1369c51f3afef3867ecdd197a5373ec6c97571c061a898e7735cc86d" Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.583504 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"34ec16ac1369c51f3afef3867ecdd197a5373ec6c97571c061a898e7735cc86d"} err="failed to get container status \"34ec16ac1369c51f3afef3867ecdd197a5373ec6c97571c061a898e7735cc86d\": rpc error: code = NotFound desc = could not find container \"34ec16ac1369c51f3afef3867ecdd197a5373ec6c97571c061a898e7735cc86d\": container with ID starting with 34ec16ac1369c51f3afef3867ecdd197a5373ec6c97571c061a898e7735cc86d not found: ID does not exist" Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.583549 4894 scope.go:117] "RemoveContainer" containerID="6c59556d88bde064c2bd4442639bae5d73c2a1a77f784aacf885c144dab00646" Dec 09 16:32:50 crc kubenswrapper[4894]: E1209 16:32:50.583898 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c59556d88bde064c2bd4442639bae5d73c2a1a77f784aacf885c144dab00646\": container with ID starting with 6c59556d88bde064c2bd4442639bae5d73c2a1a77f784aacf885c144dab00646 not found: ID does not exist" containerID="6c59556d88bde064c2bd4442639bae5d73c2a1a77f784aacf885c144dab00646" Dec 09 16:32:50 crc kubenswrapper[4894]: I1209 16:32:50.583923 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c59556d88bde064c2bd4442639bae5d73c2a1a77f784aacf885c144dab00646"} err="failed to get container status \"6c59556d88bde064c2bd4442639bae5d73c2a1a77f784aacf885c144dab00646\": rpc error: code = NotFound desc = could not find container \"6c59556d88bde064c2bd4442639bae5d73c2a1a77f784aacf885c144dab00646\": container with ID starting with 6c59556d88bde064c2bd4442639bae5d73c2a1a77f784aacf885c144dab00646 not found: ID does not exist" Dec 09 16:32:52 crc kubenswrapper[4894]: I1209 16:32:52.120791 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="76ae52d7-b36a-4967-9c30-6c3fc5c9e532" path="/var/lib/kubelet/pods/76ae52d7-b36a-4967-9c30-6c3fc5c9e532/volumes" Dec 09 16:32:55 crc kubenswrapper[4894]: I1209 16:32:55.107433 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:32:55 crc kubenswrapper[4894]: E1209 16:32:55.108379 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:32:58 crc kubenswrapper[4894]: I1209 16:32:58.257913 4894 scope.go:117] "RemoveContainer" containerID="2047006fc445345d89f058d9944e8a8d61a4d649ec8743ceee1c32242849733f" Dec 09 16:32:58 crc kubenswrapper[4894]: I1209 16:32:58.302240 4894 scope.go:117] "RemoveContainer" containerID="d48676851ecbcfe69d01635612289e472766580df9d455098ee694b7e2973346" Dec 09 16:32:58 crc kubenswrapper[4894]: I1209 16:32:58.329345 4894 scope.go:117] "RemoveContainer" containerID="a628f9d9624018f1fb0884d8760be48fb582b7fab403659d4ccee6b962734e9a" Dec 09 16:33:08 crc kubenswrapper[4894]: I1209 16:33:08.107294 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:33:08 crc kubenswrapper[4894]: E1209 16:33:08.108612 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:33:20 crc kubenswrapper[4894]: I1209 16:33:20.191079 4894 generic.go:334] "Generic (PLEG): container finished" podID="4a49e9ae-60a1-4165-ae4d-c877537d8825" containerID="de3d7b6989d294b0e7fac090792ffd6f5e333ed38a4800fdfce3729860b2f126" exitCode=123 Dec 09 16:33:20 crc kubenswrapper[4894]: I1209 16:33:20.191153 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"4a49e9ae-60a1-4165-ae4d-c877537d8825","Type":"ContainerDied","Data":"de3d7b6989d294b0e7fac090792ffd6f5e333ed38a4800fdfce3729860b2f126"} Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.771895 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.819859 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4a49e9ae-60a1-4165-ae4d-c877537d8825-config-data\") pod \"4a49e9ae-60a1-4165-ae4d-c877537d8825\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.819943 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xvlqh\" (UniqueName: \"kubernetes.io/projected/4a49e9ae-60a1-4165-ae4d-c877537d8825-kube-api-access-xvlqh\") pod \"4a49e9ae-60a1-4165-ae4d-c877537d8825\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.820041 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"4a49e9ae-60a1-4165-ae4d-c877537d8825\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.820801 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-ssh-key\") pod \"4a49e9ae-60a1-4165-ae4d-c877537d8825\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.820923 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a49e9ae-60a1-4165-ae4d-c877537d8825-config-data" (OuterVolumeSpecName: "config-data") pod "4a49e9ae-60a1-4165-ae4d-c877537d8825" (UID: "4a49e9ae-60a1-4165-ae4d-c877537d8825"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.820954 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-ca-certs\") pod \"4a49e9ae-60a1-4165-ae4d-c877537d8825\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.821040 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4a49e9ae-60a1-4165-ae4d-c877537d8825-openstack-config\") pod \"4a49e9ae-60a1-4165-ae4d-c877537d8825\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.821074 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-openstack-config-secret\") pod \"4a49e9ae-60a1-4165-ae4d-c877537d8825\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.821106 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-ceph\") pod \"4a49e9ae-60a1-4165-ae4d-c877537d8825\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.821161 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4a49e9ae-60a1-4165-ae4d-c877537d8825-test-operator-ephemeral-temporary\") pod \"4a49e9ae-60a1-4165-ae4d-c877537d8825\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.821222 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4a49e9ae-60a1-4165-ae4d-c877537d8825-test-operator-ephemeral-workdir\") pod \"4a49e9ae-60a1-4165-ae4d-c877537d8825\" (UID: \"4a49e9ae-60a1-4165-ae4d-c877537d8825\") " Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.822134 4894 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/4a49e9ae-60a1-4165-ae4d-c877537d8825-config-data\") on node \"crc\" DevicePath \"\"" Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.823401 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a49e9ae-60a1-4165-ae4d-c877537d8825-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "4a49e9ae-60a1-4165-ae4d-c877537d8825" (UID: "4a49e9ae-60a1-4165-ae4d-c877537d8825"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.842130 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4a49e9ae-60a1-4165-ae4d-c877537d8825-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "4a49e9ae-60a1-4165-ae4d-c877537d8825" (UID: "4a49e9ae-60a1-4165-ae4d-c877537d8825"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.842803 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-ceph" (OuterVolumeSpecName: "ceph") pod "4a49e9ae-60a1-4165-ae4d-c877537d8825" (UID: "4a49e9ae-60a1-4165-ae4d-c877537d8825"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.842976 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "test-operator-logs") pod "4a49e9ae-60a1-4165-ae4d-c877537d8825" (UID: "4a49e9ae-60a1-4165-ae4d-c877537d8825"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.842977 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4a49e9ae-60a1-4165-ae4d-c877537d8825-kube-api-access-xvlqh" (OuterVolumeSpecName: "kube-api-access-xvlqh") pod "4a49e9ae-60a1-4165-ae4d-c877537d8825" (UID: "4a49e9ae-60a1-4165-ae4d-c877537d8825"). InnerVolumeSpecName "kube-api-access-xvlqh". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.855332 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "4a49e9ae-60a1-4165-ae4d-c877537d8825" (UID: "4a49e9ae-60a1-4165-ae4d-c877537d8825"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.864394 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "4a49e9ae-60a1-4165-ae4d-c877537d8825" (UID: "4a49e9ae-60a1-4165-ae4d-c877537d8825"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.882564 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-ssh-key" (OuterVolumeSpecName: "ssh-key") pod "4a49e9ae-60a1-4165-ae4d-c877537d8825" (UID: "4a49e9ae-60a1-4165-ae4d-c877537d8825"). InnerVolumeSpecName "ssh-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.918914 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4a49e9ae-60a1-4165-ae4d-c877537d8825-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "4a49e9ae-60a1-4165-ae4d-c877537d8825" (UID: "4a49e9ae-60a1-4165-ae4d-c877537d8825"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.927992 4894 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.928033 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/4a49e9ae-60a1-4165-ae4d-c877537d8825-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.928051 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.928066 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.928080 4894 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/4a49e9ae-60a1-4165-ae4d-c877537d8825-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.928093 4894 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/4a49e9ae-60a1-4165-ae4d-c877537d8825-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.928107 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xvlqh\" (UniqueName: \"kubernetes.io/projected/4a49e9ae-60a1-4165-ae4d-c877537d8825-kube-api-access-xvlqh\") on node \"crc\" DevicePath \"\"" Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.928143 4894 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.928157 4894 reconciler_common.go:293] "Volume detached for volume \"ssh-key\" (UniqueName: \"kubernetes.io/secret/4a49e9ae-60a1-4165-ae4d-c877537d8825-ssh-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:33:21 crc kubenswrapper[4894]: I1209 16:33:21.956779 4894 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Dec 09 16:33:22 crc kubenswrapper[4894]: I1209 16:33:22.030047 4894 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Dec 09 16:33:22 crc kubenswrapper[4894]: I1209 16:33:22.106965 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:33:22 crc kubenswrapper[4894]: E1209 16:33:22.107562 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:33:22 crc kubenswrapper[4894]: I1209 16:33:22.232072 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tempest-tests-tempest-s01-single-test" event={"ID":"4a49e9ae-60a1-4165-ae4d-c877537d8825","Type":"ContainerDied","Data":"16d24c88b9e0f80169873c3c2bdbada7e879182674ed73d50a17afb842091c4b"} Dec 09 16:33:22 crc kubenswrapper[4894]: I1209 16:33:22.232127 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="16d24c88b9e0f80169873c3c2bdbada7e879182674ed73d50a17afb842091c4b" Dec 09 16:33:22 crc kubenswrapper[4894]: I1209 16:33:22.232148 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tempest-tests-tempest-s01-single-test" Dec 09 16:33:34 crc kubenswrapper[4894]: I1209 16:33:34.482629 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 09 16:33:34 crc kubenswrapper[4894]: E1209 16:33:34.484544 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76ae52d7-b36a-4967-9c30-6c3fc5c9e532" containerName="extract-content" Dec 09 16:33:34 crc kubenswrapper[4894]: I1209 16:33:34.484580 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="76ae52d7-b36a-4967-9c30-6c3fc5c9e532" containerName="extract-content" Dec 09 16:33:34 crc kubenswrapper[4894]: E1209 16:33:34.484607 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76ae52d7-b36a-4967-9c30-6c3fc5c9e532" containerName="registry-server" Dec 09 16:33:34 crc kubenswrapper[4894]: I1209 16:33:34.484618 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="76ae52d7-b36a-4967-9c30-6c3fc5c9e532" containerName="registry-server" Dec 09 16:33:34 crc kubenswrapper[4894]: E1209 16:33:34.484666 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4a49e9ae-60a1-4165-ae4d-c877537d8825" containerName="tempest-tests-tempest-tests-runner" Dec 09 16:33:34 crc kubenswrapper[4894]: I1209 16:33:34.484679 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="4a49e9ae-60a1-4165-ae4d-c877537d8825" containerName="tempest-tests-tempest-tests-runner" Dec 09 16:33:34 crc kubenswrapper[4894]: E1209 16:33:34.484709 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="76ae52d7-b36a-4967-9c30-6c3fc5c9e532" containerName="extract-utilities" Dec 09 16:33:34 crc kubenswrapper[4894]: I1209 16:33:34.484721 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="76ae52d7-b36a-4967-9c30-6c3fc5c9e532" containerName="extract-utilities" Dec 09 16:33:34 crc kubenswrapper[4894]: I1209 16:33:34.485054 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="4a49e9ae-60a1-4165-ae4d-c877537d8825" containerName="tempest-tests-tempest-tests-runner" Dec 09 16:33:34 crc kubenswrapper[4894]: I1209 16:33:34.485085 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="76ae52d7-b36a-4967-9c30-6c3fc5c9e532" containerName="registry-server" Dec 09 16:33:34 crc kubenswrapper[4894]: I1209 16:33:34.486347 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 09 16:33:34 crc kubenswrapper[4894]: I1209 16:33:34.488963 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"default-dockercfg-fxksq" Dec 09 16:33:34 crc kubenswrapper[4894]: I1209 16:33:34.492150 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 09 16:33:34 crc kubenswrapper[4894]: I1209 16:33:34.594124 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j579j\" (UniqueName: \"kubernetes.io/projected/5ecfe281-8bc0-42c3-af32-4b4b4be26a88-kube-api-access-j579j\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5ecfe281-8bc0-42c3-af32-4b4b4be26a88\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 09 16:33:34 crc kubenswrapper[4894]: I1209 16:33:34.594184 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5ecfe281-8bc0-42c3-af32-4b4b4be26a88\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 09 16:33:34 crc kubenswrapper[4894]: I1209 16:33:34.696496 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j579j\" (UniqueName: \"kubernetes.io/projected/5ecfe281-8bc0-42c3-af32-4b4b4be26a88-kube-api-access-j579j\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5ecfe281-8bc0-42c3-af32-4b4b4be26a88\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 09 16:33:34 crc kubenswrapper[4894]: I1209 16:33:34.696534 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5ecfe281-8bc0-42c3-af32-4b4b4be26a88\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 09 16:33:34 crc kubenswrapper[4894]: I1209 16:33:34.697134 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5ecfe281-8bc0-42c3-af32-4b4b4be26a88\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 09 16:33:34 crc kubenswrapper[4894]: I1209 16:33:34.718973 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j579j\" (UniqueName: \"kubernetes.io/projected/5ecfe281-8bc0-42c3-af32-4b4b4be26a88-kube-api-access-j579j\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5ecfe281-8bc0-42c3-af32-4b4b4be26a88\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 09 16:33:34 crc kubenswrapper[4894]: I1209 16:33:34.726371 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"test-operator-logs-pod-tempest-tempest-tests-tempest\" (UID: \"5ecfe281-8bc0-42c3-af32-4b4b4be26a88\") " pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 09 16:33:34 crc kubenswrapper[4894]: I1209 16:33:34.809814 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" Dec 09 16:33:35 crc kubenswrapper[4894]: I1209 16:33:35.107331 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:33:35 crc kubenswrapper[4894]: E1209 16:33:35.108263 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:33:35 crc kubenswrapper[4894]: I1209 16:33:35.305310 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tempest-tempest-tests-tempest"] Dec 09 16:33:35 crc kubenswrapper[4894]: I1209 16:33:35.376036 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"5ecfe281-8bc0-42c3-af32-4b4b4be26a88","Type":"ContainerStarted","Data":"cea6600a561fcf6312fe2761f49308a64eff26adb6967f2b4e5e137fb95cb82f"} Dec 09 16:33:37 crc kubenswrapper[4894]: I1209 16:33:37.401163 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" event={"ID":"5ecfe281-8bc0-42c3-af32-4b4b4be26a88","Type":"ContainerStarted","Data":"ba76a82cb3ad7bd2eea8cb10a1f9a7fac266d20364e956372a5c735a90b3fcf6"} Dec 09 16:33:37 crc kubenswrapper[4894]: I1209 16:33:37.434354 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tempest-tempest-tests-tempest" podStartSLOduration=1.774594412 podStartE2EDuration="3.43432294s" podCreationTimestamp="2025-12-09 16:33:34 +0000 UTC" firstStartedPulling="2025-12-09 16:33:35.312238549 +0000 UTC m=+3709.631449218" lastFinishedPulling="2025-12-09 16:33:36.971967087 +0000 UTC m=+3711.291177746" observedRunningTime="2025-12-09 16:33:37.417580704 +0000 UTC m=+3711.736791423" watchObservedRunningTime="2025-12-09 16:33:37.43432294 +0000 UTC m=+3711.753533639" Dec 09 16:33:49 crc kubenswrapper[4894]: I1209 16:33:49.106521 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:33:49 crc kubenswrapper[4894]: E1209 16:33:49.107331 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:33:52 crc kubenswrapper[4894]: I1209 16:33:52.887930 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tobiko-tests-tobiko-s00-podified-functional"] Dec 09 16:33:52 crc kubenswrapper[4894]: I1209 16:33:52.890129 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:52 crc kubenswrapper[4894]: I1209 16:33:52.897701 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tobiko-tests-tobikotobiko-config" Dec 09 16:33:52 crc kubenswrapper[4894]: I1209 16:33:52.901568 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"tobiko-secret" Dec 09 16:33:52 crc kubenswrapper[4894]: I1209 16:33:52.901865 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"test-operator-clouds-config" Dec 09 16:33:52 crc kubenswrapper[4894]: I1209 16:33:52.902065 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tobiko-tests-tobikotobiko-private-key" Dec 09 16:33:52 crc kubenswrapper[4894]: I1209 16:33:52.902485 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"tobiko-tests-tobikotobiko-public-key" Dec 09 16:33:52 crc kubenswrapper[4894]: I1209 16:33:52.934942 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s00-podified-functional"] Dec 09 16:33:52 crc kubenswrapper[4894]: I1209 16:33:52.955122 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:52 crc kubenswrapper[4894]: I1209 16:33:52.955198 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/f3302421-5f23-4e31-b6b7-9ea816c8fdea-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:52 crc kubenswrapper[4894]: I1209 16:33:52.955321 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-kubeconfig\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:52 crc kubenswrapper[4894]: I1209 16:33:52.955379 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-tobiko-public-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:52 crc kubenswrapper[4894]: I1209 16:33:52.955437 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:52 crc kubenswrapper[4894]: I1209 16:33:52.955538 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-openstack-config-secret\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:52 crc kubenswrapper[4894]: I1209 16:33:52.955719 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-tobiko-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:52 crc kubenswrapper[4894]: I1209 16:33:52.955918 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/f3302421-5f23-4e31-b6b7-9ea816c8fdea-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:52 crc kubenswrapper[4894]: I1209 16:33:52.956006 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-ca-certs\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:52 crc kubenswrapper[4894]: I1209 16:33:52.956095 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-ceph\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:52 crc kubenswrapper[4894]: I1209 16:33:52.956159 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-tobiko-private-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:52 crc kubenswrapper[4894]: I1209 16:33:52.956221 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5qr87\" (UniqueName: \"kubernetes.io/projected/f3302421-5f23-4e31-b6b7-9ea816c8fdea-kube-api-access-5qr87\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.057970 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-kubeconfig\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.058010 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-tobiko-public-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.058033 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.058068 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-openstack-config-secret\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.058135 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-tobiko-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.058177 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/f3302421-5f23-4e31-b6b7-9ea816c8fdea-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.058218 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-ca-certs\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.058254 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-ceph\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.058284 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-tobiko-private-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.058317 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5qr87\" (UniqueName: \"kubernetes.io/projected/f3302421-5f23-4e31-b6b7-9ea816c8fdea-kube-api-access-5qr87\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.058345 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.058387 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/f3302421-5f23-4e31-b6b7-9ea816c8fdea-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.058860 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/f3302421-5f23-4e31-b6b7-9ea816c8fdea-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.059355 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-tobiko-public-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.060037 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/f3302421-5f23-4e31-b6b7-9ea816c8fdea-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.059808 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.060998 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-tobiko-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.059505 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-tobiko-private-key\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.062619 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.065018 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-ceph\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.066081 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-openstack-config-secret\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.068740 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-kubeconfig\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.080775 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-ca-certs\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.082669 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5qr87\" (UniqueName: \"kubernetes.io/projected/f3302421-5f23-4e31-b6b7-9ea816c8fdea-kube-api-access-5qr87\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.107513 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tobiko-tests-tobiko-s00-podified-functional\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.235092 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:33:53 crc kubenswrapper[4894]: I1209 16:33:53.832292 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s00-podified-functional"] Dec 09 16:33:53 crc kubenswrapper[4894]: W1209 16:33:53.840316 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3302421_5f23_4e31_b6b7_9ea816c8fdea.slice/crio-5d1b31b555a825f3f20f071f873f973d346e15cb2c19134db8c67e8209f29d2d WatchSource:0}: Error finding container 5d1b31b555a825f3f20f071f873f973d346e15cb2c19134db8c67e8209f29d2d: Status 404 returned error can't find the container with id 5d1b31b555a825f3f20f071f873f973d346e15cb2c19134db8c67e8209f29d2d Dec 09 16:33:54 crc kubenswrapper[4894]: I1209 16:33:54.585319 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"f3302421-5f23-4e31-b6b7-9ea816c8fdea","Type":"ContainerStarted","Data":"5d1b31b555a825f3f20f071f873f973d346e15cb2c19134db8c67e8209f29d2d"} Dec 09 16:34:02 crc kubenswrapper[4894]: I1209 16:34:02.113895 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:34:02 crc kubenswrapper[4894]: E1209 16:34:02.114607 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:34:15 crc kubenswrapper[4894]: I1209 16:34:15.107480 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:34:15 crc kubenswrapper[4894]: E1209 16:34:15.108775 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:34:16 crc kubenswrapper[4894]: E1209 16:34:16.214883 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-tobiko:current-podified" Dec 09 16:34:16 crc kubenswrapper[4894]: E1209 16:34:16.215487 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:tobiko-tests-tobiko,Image:quay.io/podified-antelope-centos9/openstack-tobiko:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TOBIKO_DEBUG_MODE,Value:false,ValueFrom:nil,},EnvVar{Name:TOBIKO_KEYS_FOLDER,Value:/etc/test_operator,ValueFrom:nil,},EnvVar{Name:TOBIKO_LOGS_DIR_NAME,Value:tobiko-tests-tobiko-s00-podified-functional,ValueFrom:nil,},EnvVar{Name:TOBIKO_PREVENT_CREATE,Value:,ValueFrom:nil,},EnvVar{Name:TOBIKO_PYTEST_ADDOPTS,Value:,ValueFrom:nil,},EnvVar{Name:TOBIKO_TESTENV,Value:functional -- tobiko/tests/functional/podified/test_topology.py,ValueFrom:nil,},EnvVar{Name:TOBIKO_VERSION,Value:master,ValueFrom:nil,},EnvVar{Name:TOX_NUM_PROCESSES,Value:2,ValueFrom:nil,},EnvVar{Name:USE_EXTERNAL_FILES,Value:True,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{8 0} {} 8 DecimalSI},memory: {{8589934592 0} {} BinarySI},},Requests:ResourceList{cpu: {{4 0} {} 4 DecimalSI},memory: {{4294967296 0} {} 4Gi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/tobiko,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/tobiko/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/var/lib/tobiko/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tobiko-config,ReadOnly:false,MountPath:/etc/tobiko/tobiko.conf,SubPath:tobiko.conf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ca-bundle.trust.crt,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tobiko-private-key,ReadOnly:true,MountPath:/etc/test_operator/id_ecdsa,SubPath:id_ecdsa,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:tobiko-public-key,ReadOnly:true,MountPath:/etc/test_operator/id_ecdsa.pub,SubPath:id_ecdsa.pub,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kubeconfig,ReadOnly:true,MountPath:/var/lib/tobiko/.kube/config,SubPath:config,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5qr87,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN NET_RAW],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42495,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42495,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod tobiko-tests-tobiko-s00-podified-functional_openstack(f3302421-5f23-4e31-b6b7-9ea816c8fdea): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 16:34:16 crc kubenswrapper[4894]: E1209 16:34:16.216831 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tobiko-tests-tobiko\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" podUID="f3302421-5f23-4e31-b6b7-9ea816c8fdea" Dec 09 16:34:16 crc kubenswrapper[4894]: E1209 16:34:16.824592 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"tobiko-tests-tobiko\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-tobiko:current-podified\\\"\"" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" podUID="f3302421-5f23-4e31-b6b7-9ea816c8fdea" Dec 09 16:34:27 crc kubenswrapper[4894]: I1209 16:34:27.106169 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:34:27 crc kubenswrapper[4894]: E1209 16:34:27.107052 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:34:33 crc kubenswrapper[4894]: I1209 16:34:33.041390 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"f3302421-5f23-4e31-b6b7-9ea816c8fdea","Type":"ContainerStarted","Data":"ba2a2e6b89c697c26dd1589e6a7f83d9603b0563e28c253ee868aa422d45aaa1"} Dec 09 16:34:33 crc kubenswrapper[4894]: I1209 16:34:33.066945 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" podStartSLOduration=4.047406225 podStartE2EDuration="42.066916083s" podCreationTimestamp="2025-12-09 16:33:51 +0000 UTC" firstStartedPulling="2025-12-09 16:33:53.842405422 +0000 UTC m=+3728.161616131" lastFinishedPulling="2025-12-09 16:34:31.86191532 +0000 UTC m=+3766.181125989" observedRunningTime="2025-12-09 16:34:33.062142293 +0000 UTC m=+3767.381352982" watchObservedRunningTime="2025-12-09 16:34:33.066916083 +0000 UTC m=+3767.386126792" Dec 09 16:34:38 crc kubenswrapper[4894]: I1209 16:34:38.110593 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:34:38 crc kubenswrapper[4894]: E1209 16:34:38.111434 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:34:52 crc kubenswrapper[4894]: I1209 16:34:52.106858 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:34:52 crc kubenswrapper[4894]: E1209 16:34:52.107613 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:35:06 crc kubenswrapper[4894]: I1209 16:35:06.114868 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:35:06 crc kubenswrapper[4894]: E1209 16:35:06.115974 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:35:21 crc kubenswrapper[4894]: I1209 16:35:21.106482 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:35:21 crc kubenswrapper[4894]: E1209 16:35:21.107294 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:35:33 crc kubenswrapper[4894]: I1209 16:35:33.318198 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-tgkbq"] Dec 09 16:35:33 crc kubenswrapper[4894]: I1209 16:35:33.321498 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tgkbq" Dec 09 16:35:33 crc kubenswrapper[4894]: I1209 16:35:33.342033 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tgkbq"] Dec 09 16:35:33 crc kubenswrapper[4894]: I1209 16:35:33.397083 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1140abe-2fbd-449f-be06-e6581f5757c2-utilities\") pod \"certified-operators-tgkbq\" (UID: \"a1140abe-2fbd-449f-be06-e6581f5757c2\") " pod="openshift-marketplace/certified-operators-tgkbq" Dec 09 16:35:33 crc kubenswrapper[4894]: I1209 16:35:33.397145 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1140abe-2fbd-449f-be06-e6581f5757c2-catalog-content\") pod \"certified-operators-tgkbq\" (UID: \"a1140abe-2fbd-449f-be06-e6581f5757c2\") " pod="openshift-marketplace/certified-operators-tgkbq" Dec 09 16:35:33 crc kubenswrapper[4894]: I1209 16:35:33.397251 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmws9\" (UniqueName: \"kubernetes.io/projected/a1140abe-2fbd-449f-be06-e6581f5757c2-kube-api-access-zmws9\") pod \"certified-operators-tgkbq\" (UID: \"a1140abe-2fbd-449f-be06-e6581f5757c2\") " pod="openshift-marketplace/certified-operators-tgkbq" Dec 09 16:35:33 crc kubenswrapper[4894]: I1209 16:35:33.498866 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1140abe-2fbd-449f-be06-e6581f5757c2-utilities\") pod \"certified-operators-tgkbq\" (UID: \"a1140abe-2fbd-449f-be06-e6581f5757c2\") " pod="openshift-marketplace/certified-operators-tgkbq" Dec 09 16:35:33 crc kubenswrapper[4894]: I1209 16:35:33.498930 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1140abe-2fbd-449f-be06-e6581f5757c2-catalog-content\") pod \"certified-operators-tgkbq\" (UID: \"a1140abe-2fbd-449f-be06-e6581f5757c2\") " pod="openshift-marketplace/certified-operators-tgkbq" Dec 09 16:35:33 crc kubenswrapper[4894]: I1209 16:35:33.499038 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmws9\" (UniqueName: \"kubernetes.io/projected/a1140abe-2fbd-449f-be06-e6581f5757c2-kube-api-access-zmws9\") pod \"certified-operators-tgkbq\" (UID: \"a1140abe-2fbd-449f-be06-e6581f5757c2\") " pod="openshift-marketplace/certified-operators-tgkbq" Dec 09 16:35:33 crc kubenswrapper[4894]: I1209 16:35:33.499448 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1140abe-2fbd-449f-be06-e6581f5757c2-utilities\") pod \"certified-operators-tgkbq\" (UID: \"a1140abe-2fbd-449f-be06-e6581f5757c2\") " pod="openshift-marketplace/certified-operators-tgkbq" Dec 09 16:35:33 crc kubenswrapper[4894]: I1209 16:35:33.499822 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1140abe-2fbd-449f-be06-e6581f5757c2-catalog-content\") pod \"certified-operators-tgkbq\" (UID: \"a1140abe-2fbd-449f-be06-e6581f5757c2\") " pod="openshift-marketplace/certified-operators-tgkbq" Dec 09 16:35:33 crc kubenswrapper[4894]: I1209 16:35:33.518503 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmws9\" (UniqueName: \"kubernetes.io/projected/a1140abe-2fbd-449f-be06-e6581f5757c2-kube-api-access-zmws9\") pod \"certified-operators-tgkbq\" (UID: \"a1140abe-2fbd-449f-be06-e6581f5757c2\") " pod="openshift-marketplace/certified-operators-tgkbq" Dec 09 16:35:33 crc kubenswrapper[4894]: I1209 16:35:33.647369 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tgkbq" Dec 09 16:35:34 crc kubenswrapper[4894]: I1209 16:35:34.048358 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-tgkbq"] Dec 09 16:35:34 crc kubenswrapper[4894]: I1209 16:35:34.110538 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:35:34 crc kubenswrapper[4894]: E1209 16:35:34.110881 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:35:35 crc kubenswrapper[4894]: I1209 16:35:35.018274 4894 generic.go:334] "Generic (PLEG): container finished" podID="a1140abe-2fbd-449f-be06-e6581f5757c2" containerID="cee876fcbf9f2dfbc38f3d799c099343526329041a3351d89bbd85824ec69c3b" exitCode=0 Dec 09 16:35:35 crc kubenswrapper[4894]: I1209 16:35:35.018383 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tgkbq" event={"ID":"a1140abe-2fbd-449f-be06-e6581f5757c2","Type":"ContainerDied","Data":"cee876fcbf9f2dfbc38f3d799c099343526329041a3351d89bbd85824ec69c3b"} Dec 09 16:35:35 crc kubenswrapper[4894]: I1209 16:35:35.018563 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tgkbq" event={"ID":"a1140abe-2fbd-449f-be06-e6581f5757c2","Type":"ContainerStarted","Data":"f1e734e04175d1d10427d7c3f2f4c86f525510f8d74184ef3a082146e6e06993"} Dec 09 16:35:39 crc kubenswrapper[4894]: I1209 16:35:39.062488 4894 generic.go:334] "Generic (PLEG): container finished" podID="a1140abe-2fbd-449f-be06-e6581f5757c2" containerID="a3c05b7c6413b48fff30b2f32adc2c3eeda23f0217b5ad1f9aa48f301b989535" exitCode=0 Dec 09 16:35:39 crc kubenswrapper[4894]: I1209 16:35:39.062564 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tgkbq" event={"ID":"a1140abe-2fbd-449f-be06-e6581f5757c2","Type":"ContainerDied","Data":"a3c05b7c6413b48fff30b2f32adc2c3eeda23f0217b5ad1f9aa48f301b989535"} Dec 09 16:35:41 crc kubenswrapper[4894]: I1209 16:35:41.082574 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tgkbq" event={"ID":"a1140abe-2fbd-449f-be06-e6581f5757c2","Type":"ContainerStarted","Data":"6c1f0e5a42b9bb83e8d161458cc686adc5fa126f2bd6de1041e6790affec7258"} Dec 09 16:35:41 crc kubenswrapper[4894]: I1209 16:35:41.117750 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-tgkbq" podStartSLOduration=2.388634737 podStartE2EDuration="8.117716698s" podCreationTimestamp="2025-12-09 16:35:33 +0000 UTC" firstStartedPulling="2025-12-09 16:35:35.021072474 +0000 UTC m=+3829.340283183" lastFinishedPulling="2025-12-09 16:35:40.750154475 +0000 UTC m=+3835.069365144" observedRunningTime="2025-12-09 16:35:41.104684273 +0000 UTC m=+3835.423894962" watchObservedRunningTime="2025-12-09 16:35:41.117716698 +0000 UTC m=+3835.436927387" Dec 09 16:35:43 crc kubenswrapper[4894]: I1209 16:35:43.108164 4894 generic.go:334] "Generic (PLEG): container finished" podID="f3302421-5f23-4e31-b6b7-9ea816c8fdea" containerID="ba2a2e6b89c697c26dd1589e6a7f83d9603b0563e28c253ee868aa422d45aaa1" exitCode=0 Dec 09 16:35:43 crc kubenswrapper[4894]: I1209 16:35:43.108659 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"f3302421-5f23-4e31-b6b7-9ea816c8fdea","Type":"ContainerDied","Data":"ba2a2e6b89c697c26dd1589e6a7f83d9603b0563e28c253ee868aa422d45aaa1"} Dec 09 16:35:43 crc kubenswrapper[4894]: I1209 16:35:43.647533 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-tgkbq" Dec 09 16:35:43 crc kubenswrapper[4894]: I1209 16:35:43.647581 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-tgkbq" Dec 09 16:35:43 crc kubenswrapper[4894]: I1209 16:35:43.693228 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-tgkbq" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.599254 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.665583 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/tobiko-tests-tobiko-s01-sanity"] Dec 09 16:35:44 crc kubenswrapper[4894]: E1209 16:35:44.666049 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3302421-5f23-4e31-b6b7-9ea816c8fdea" containerName="tobiko-tests-tobiko" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.666068 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3302421-5f23-4e31-b6b7-9ea816c8fdea" containerName="tobiko-tests-tobiko" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.666258 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3302421-5f23-4e31-b6b7-9ea816c8fdea" containerName="tobiko-tests-tobiko" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.666941 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.693092 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s01-sanity"] Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.736781 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-tobiko-config\") pod \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.738016 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5qr87\" (UniqueName: \"kubernetes.io/projected/f3302421-5f23-4e31-b6b7-9ea816c8fdea-kube-api-access-5qr87\") pod \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.739152 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-ceph\") pod \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.739216 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/f3302421-5f23-4e31-b6b7-9ea816c8fdea-test-operator-ephemeral-temporary\") pod \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.739280 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-kubeconfig\") pod \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.739333 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-openstack-config-secret\") pod \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.739486 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-test-operator-clouds-config\") pod \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.739517 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-ca-certs\") pod \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.739598 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.739672 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-tobiko-public-key\") pod \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.739783 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/f3302421-5f23-4e31-b6b7-9ea816c8fdea-test-operator-ephemeral-workdir\") pod \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.739828 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-tobiko-private-key\") pod \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\" (UID: \"f3302421-5f23-4e31-b6b7-9ea816c8fdea\") " Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.740281 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3302421-5f23-4e31-b6b7-9ea816c8fdea-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "f3302421-5f23-4e31-b6b7-9ea816c8fdea" (UID: "f3302421-5f23-4e31-b6b7-9ea816c8fdea"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.740430 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-tobiko-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.740528 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-openstack-config-secret\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.740657 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.740928 4894 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/f3302421-5f23-4e31-b6b7-9ea816c8fdea-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.843227 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3dd4c51e-1b24-48e9-a722-80c11a2460fd-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.843470 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-tobiko-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.843524 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-openstack-config-secret\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.843551 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-kubeconfig\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.843595 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.843618 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-ceph\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.843655 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f66nv\" (UniqueName: \"kubernetes.io/projected/3dd4c51e-1b24-48e9-a722-80c11a2460fd-kube-api-access-f66nv\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.843708 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-tobiko-public-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.843729 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3dd4c51e-1b24-48e9-a722-80c11a2460fd-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.843749 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-tobiko-private-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.843770 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-ca-certs\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.844906 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-tobiko-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.845631 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-test-operator-clouds-config\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.946788 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-kubeconfig\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.946940 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-ceph\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.946988 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f66nv\" (UniqueName: \"kubernetes.io/projected/3dd4c51e-1b24-48e9-a722-80c11a2460fd-kube-api-access-f66nv\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.947067 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-tobiko-public-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.947127 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3dd4c51e-1b24-48e9-a722-80c11a2460fd-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.947190 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-tobiko-private-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.947252 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-ca-certs\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.947303 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3dd4c51e-1b24-48e9-a722-80c11a2460fd-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.948265 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3dd4c51e-1b24-48e9-a722-80c11a2460fd-test-operator-ephemeral-workdir\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.948319 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3dd4c51e-1b24-48e9-a722-80c11a2460fd-test-operator-ephemeral-temporary\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.948346 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-tobiko-private-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:44 crc kubenswrapper[4894]: I1209 16:35:44.948431 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-tobiko-public-key\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.128875 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" event={"ID":"f3302421-5f23-4e31-b6b7-9ea816c8fdea","Type":"ContainerDied","Data":"5d1b31b555a825f3f20f071f873f973d346e15cb2c19134db8c67e8209f29d2d"} Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.128938 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5d1b31b555a825f3f20f071f873f973d346e15cb2c19134db8c67e8209f29d2d" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.128953 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s00-podified-functional" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.407555 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-ceph\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.407652 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-openstack-config-secret\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.420242 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f66nv\" (UniqueName: \"kubernetes.io/projected/3dd4c51e-1b24-48e9-a722-80c11a2460fd-kube-api-access-f66nv\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.420474 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-kubeconfig\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.420874 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-ca-certs\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.504085 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "test-operator-logs") pod "f3302421-5f23-4e31-b6b7-9ea816c8fdea" (UID: "f3302421-5f23-4e31-b6b7-9ea816c8fdea"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.504183 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3302421-5f23-4e31-b6b7-9ea816c8fdea-kube-api-access-5qr87" (OuterVolumeSpecName: "kube-api-access-5qr87") pod "f3302421-5f23-4e31-b6b7-9ea816c8fdea" (UID: "f3302421-5f23-4e31-b6b7-9ea816c8fdea"). InnerVolumeSpecName "kube-api-access-5qr87". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.504806 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-ceph" (OuterVolumeSpecName: "ceph") pod "f3302421-5f23-4e31-b6b7-9ea816c8fdea" (UID: "f3302421-5f23-4e31-b6b7-9ea816c8fdea"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.505510 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-tobiko-public-key" (OuterVolumeSpecName: "tobiko-public-key") pod "f3302421-5f23-4e31-b6b7-9ea816c8fdea" (UID: "f3302421-5f23-4e31-b6b7-9ea816c8fdea"). InnerVolumeSpecName "tobiko-public-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.505912 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-tobiko-config" (OuterVolumeSpecName: "tobiko-config") pod "f3302421-5f23-4e31-b6b7-9ea816c8fdea" (UID: "f3302421-5f23-4e31-b6b7-9ea816c8fdea"). InnerVolumeSpecName "tobiko-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.509246 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "f3302421-5f23-4e31-b6b7-9ea816c8fdea" (UID: "f3302421-5f23-4e31-b6b7-9ea816c8fdea"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.509398 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-tobiko-private-key" (OuterVolumeSpecName: "tobiko-private-key") pod "f3302421-5f23-4e31-b6b7-9ea816c8fdea" (UID: "f3302421-5f23-4e31-b6b7-9ea816c8fdea"). InnerVolumeSpecName "tobiko-private-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.511443 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-kubeconfig" (OuterVolumeSpecName: "kubeconfig") pod "f3302421-5f23-4e31-b6b7-9ea816c8fdea" (UID: "f3302421-5f23-4e31-b6b7-9ea816c8fdea"). InnerVolumeSpecName "kubeconfig". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.531655 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "f3302421-5f23-4e31-b6b7-9ea816c8fdea" (UID: "f3302421-5f23-4e31-b6b7-9ea816c8fdea"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.533598 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-test-operator-clouds-config" (OuterVolumeSpecName: "test-operator-clouds-config") pod "f3302421-5f23-4e31-b6b7-9ea816c8fdea" (UID: "f3302421-5f23-4e31-b6b7-9ea816c8fdea"). InnerVolumeSpecName "test-operator-clouds-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.559606 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.559795 4894 reconciler_common.go:293] "Volume detached for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-tobiko-public-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.559808 4894 reconciler_common.go:293] "Volume detached for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-tobiko-private-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.559816 4894 reconciler_common.go:293] "Volume detached for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-tobiko-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.559825 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5qr87\" (UniqueName: \"kubernetes.io/projected/f3302421-5f23-4e31-b6b7-9ea816c8fdea-kube-api-access-5qr87\") on node \"crc\" DevicePath \"\"" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.559836 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.559848 4894 reconciler_common.go:293] "Volume detached for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-kubeconfig\") on node \"crc\" DevicePath \"\"" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.559857 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.559867 4894 reconciler_common.go:293] "Volume detached for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/f3302421-5f23-4e31-b6b7-9ea816c8fdea-test-operator-clouds-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.559875 4894 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/f3302421-5f23-4e31-b6b7-9ea816c8fdea-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.588261 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"tobiko-tests-tobiko-s01-sanity\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:45 crc kubenswrapper[4894]: I1209 16:35:45.601854 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:35:46 crc kubenswrapper[4894]: I1209 16:35:46.114088 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:35:46 crc kubenswrapper[4894]: E1209 16:35:46.121522 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:35:46 crc kubenswrapper[4894]: I1209 16:35:46.133018 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/tobiko-tests-tobiko-s01-sanity"] Dec 09 16:35:46 crc kubenswrapper[4894]: I1209 16:35:46.142090 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3302421-5f23-4e31-b6b7-9ea816c8fdea-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "f3302421-5f23-4e31-b6b7-9ea816c8fdea" (UID: "f3302421-5f23-4e31-b6b7-9ea816c8fdea"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:35:46 crc kubenswrapper[4894]: I1209 16:35:46.174734 4894 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/f3302421-5f23-4e31-b6b7-9ea816c8fdea-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 09 16:35:47 crc kubenswrapper[4894]: I1209 16:35:47.147278 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"3dd4c51e-1b24-48e9-a722-80c11a2460fd","Type":"ContainerStarted","Data":"8572fc20ed4efd6d16c60ad7c98da04cb647ff26aac0271aa26134c9ca70fc6f"} Dec 09 16:35:47 crc kubenswrapper[4894]: I1209 16:35:47.147578 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"3dd4c51e-1b24-48e9-a722-80c11a2460fd","Type":"ContainerStarted","Data":"32ae6c77da3557f1ed549340b7f0c52b5e67b25b21fb09a0089013eeeb8e89fe"} Dec 09 16:35:47 crc kubenswrapper[4894]: I1209 16:35:47.175006 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/tobiko-tests-tobiko-s01-sanity" podStartSLOduration=3.17498033 podStartE2EDuration="3.17498033s" podCreationTimestamp="2025-12-09 16:35:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-12-09 16:35:47.171601239 +0000 UTC m=+3841.490811958" watchObservedRunningTime="2025-12-09 16:35:47.17498033 +0000 UTC m=+3841.494191019" Dec 09 16:35:53 crc kubenswrapper[4894]: I1209 16:35:53.696516 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-tgkbq" Dec 09 16:35:53 crc kubenswrapper[4894]: I1209 16:35:53.756202 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tgkbq"] Dec 09 16:35:54 crc kubenswrapper[4894]: I1209 16:35:54.226141 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-tgkbq" podUID="a1140abe-2fbd-449f-be06-e6581f5757c2" containerName="registry-server" containerID="cri-o://6c1f0e5a42b9bb83e8d161458cc686adc5fa126f2bd6de1041e6790affec7258" gracePeriod=2 Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.197865 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tgkbq" Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.240008 4894 generic.go:334] "Generic (PLEG): container finished" podID="a1140abe-2fbd-449f-be06-e6581f5757c2" containerID="6c1f0e5a42b9bb83e8d161458cc686adc5fa126f2bd6de1041e6790affec7258" exitCode=0 Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.240044 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tgkbq" event={"ID":"a1140abe-2fbd-449f-be06-e6581f5757c2","Type":"ContainerDied","Data":"6c1f0e5a42b9bb83e8d161458cc686adc5fa126f2bd6de1041e6790affec7258"} Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.240065 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-tgkbq" event={"ID":"a1140abe-2fbd-449f-be06-e6581f5757c2","Type":"ContainerDied","Data":"f1e734e04175d1d10427d7c3f2f4c86f525510f8d74184ef3a082146e6e06993"} Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.240082 4894 scope.go:117] "RemoveContainer" containerID="6c1f0e5a42b9bb83e8d161458cc686adc5fa126f2bd6de1041e6790affec7258" Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.240191 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-tgkbq" Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.276353 4894 scope.go:117] "RemoveContainer" containerID="a3c05b7c6413b48fff30b2f32adc2c3eeda23f0217b5ad1f9aa48f301b989535" Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.301646 4894 scope.go:117] "RemoveContainer" containerID="cee876fcbf9f2dfbc38f3d799c099343526329041a3351d89bbd85824ec69c3b" Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.341459 4894 scope.go:117] "RemoveContainer" containerID="6c1f0e5a42b9bb83e8d161458cc686adc5fa126f2bd6de1041e6790affec7258" Dec 09 16:35:55 crc kubenswrapper[4894]: E1209 16:35:55.342025 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c1f0e5a42b9bb83e8d161458cc686adc5fa126f2bd6de1041e6790affec7258\": container with ID starting with 6c1f0e5a42b9bb83e8d161458cc686adc5fa126f2bd6de1041e6790affec7258 not found: ID does not exist" containerID="6c1f0e5a42b9bb83e8d161458cc686adc5fa126f2bd6de1041e6790affec7258" Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.342061 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c1f0e5a42b9bb83e8d161458cc686adc5fa126f2bd6de1041e6790affec7258"} err="failed to get container status \"6c1f0e5a42b9bb83e8d161458cc686adc5fa126f2bd6de1041e6790affec7258\": rpc error: code = NotFound desc = could not find container \"6c1f0e5a42b9bb83e8d161458cc686adc5fa126f2bd6de1041e6790affec7258\": container with ID starting with 6c1f0e5a42b9bb83e8d161458cc686adc5fa126f2bd6de1041e6790affec7258 not found: ID does not exist" Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.342085 4894 scope.go:117] "RemoveContainer" containerID="a3c05b7c6413b48fff30b2f32adc2c3eeda23f0217b5ad1f9aa48f301b989535" Dec 09 16:35:55 crc kubenswrapper[4894]: E1209 16:35:55.342437 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a3c05b7c6413b48fff30b2f32adc2c3eeda23f0217b5ad1f9aa48f301b989535\": container with ID starting with a3c05b7c6413b48fff30b2f32adc2c3eeda23f0217b5ad1f9aa48f301b989535 not found: ID does not exist" containerID="a3c05b7c6413b48fff30b2f32adc2c3eeda23f0217b5ad1f9aa48f301b989535" Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.342464 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a3c05b7c6413b48fff30b2f32adc2c3eeda23f0217b5ad1f9aa48f301b989535"} err="failed to get container status \"a3c05b7c6413b48fff30b2f32adc2c3eeda23f0217b5ad1f9aa48f301b989535\": rpc error: code = NotFound desc = could not find container \"a3c05b7c6413b48fff30b2f32adc2c3eeda23f0217b5ad1f9aa48f301b989535\": container with ID starting with a3c05b7c6413b48fff30b2f32adc2c3eeda23f0217b5ad1f9aa48f301b989535 not found: ID does not exist" Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.342480 4894 scope.go:117] "RemoveContainer" containerID="cee876fcbf9f2dfbc38f3d799c099343526329041a3351d89bbd85824ec69c3b" Dec 09 16:35:55 crc kubenswrapper[4894]: E1209 16:35:55.342835 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cee876fcbf9f2dfbc38f3d799c099343526329041a3351d89bbd85824ec69c3b\": container with ID starting with cee876fcbf9f2dfbc38f3d799c099343526329041a3351d89bbd85824ec69c3b not found: ID does not exist" containerID="cee876fcbf9f2dfbc38f3d799c099343526329041a3351d89bbd85824ec69c3b" Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.342873 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cee876fcbf9f2dfbc38f3d799c099343526329041a3351d89bbd85824ec69c3b"} err="failed to get container status \"cee876fcbf9f2dfbc38f3d799c099343526329041a3351d89bbd85824ec69c3b\": rpc error: code = NotFound desc = could not find container \"cee876fcbf9f2dfbc38f3d799c099343526329041a3351d89bbd85824ec69c3b\": container with ID starting with cee876fcbf9f2dfbc38f3d799c099343526329041a3351d89bbd85824ec69c3b not found: ID does not exist" Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.365080 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1140abe-2fbd-449f-be06-e6581f5757c2-utilities\") pod \"a1140abe-2fbd-449f-be06-e6581f5757c2\" (UID: \"a1140abe-2fbd-449f-be06-e6581f5757c2\") " Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.365348 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zmws9\" (UniqueName: \"kubernetes.io/projected/a1140abe-2fbd-449f-be06-e6581f5757c2-kube-api-access-zmws9\") pod \"a1140abe-2fbd-449f-be06-e6581f5757c2\" (UID: \"a1140abe-2fbd-449f-be06-e6581f5757c2\") " Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.365556 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1140abe-2fbd-449f-be06-e6581f5757c2-catalog-content\") pod \"a1140abe-2fbd-449f-be06-e6581f5757c2\" (UID: \"a1140abe-2fbd-449f-be06-e6581f5757c2\") " Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.366352 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1140abe-2fbd-449f-be06-e6581f5757c2-utilities" (OuterVolumeSpecName: "utilities") pod "a1140abe-2fbd-449f-be06-e6581f5757c2" (UID: "a1140abe-2fbd-449f-be06-e6581f5757c2"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.375944 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a1140abe-2fbd-449f-be06-e6581f5757c2-kube-api-access-zmws9" (OuterVolumeSpecName: "kube-api-access-zmws9") pod "a1140abe-2fbd-449f-be06-e6581f5757c2" (UID: "a1140abe-2fbd-449f-be06-e6581f5757c2"). InnerVolumeSpecName "kube-api-access-zmws9". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.426096 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a1140abe-2fbd-449f-be06-e6581f5757c2-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a1140abe-2fbd-449f-be06-e6581f5757c2" (UID: "a1140abe-2fbd-449f-be06-e6581f5757c2"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.468510 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zmws9\" (UniqueName: \"kubernetes.io/projected/a1140abe-2fbd-449f-be06-e6581f5757c2-kube-api-access-zmws9\") on node \"crc\" DevicePath \"\"" Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.468542 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a1140abe-2fbd-449f-be06-e6581f5757c2-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.468554 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a1140abe-2fbd-449f-be06-e6581f5757c2-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.597285 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-tgkbq"] Dec 09 16:35:55 crc kubenswrapper[4894]: I1209 16:35:55.614136 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-tgkbq"] Dec 09 16:35:56 crc kubenswrapper[4894]: I1209 16:35:56.122783 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a1140abe-2fbd-449f-be06-e6581f5757c2" path="/var/lib/kubelet/pods/a1140abe-2fbd-449f-be06-e6581f5757c2/volumes" Dec 09 16:36:00 crc kubenswrapper[4894]: I1209 16:36:00.107870 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:36:00 crc kubenswrapper[4894]: E1209 16:36:00.109343 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:36:13 crc kubenswrapper[4894]: I1209 16:36:13.105908 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:36:13 crc kubenswrapper[4894]: E1209 16:36:13.107586 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:36:24 crc kubenswrapper[4894]: I1209 16:36:24.112889 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:36:24 crc kubenswrapper[4894]: E1209 16:36:24.113766 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:36:39 crc kubenswrapper[4894]: I1209 16:36:39.106796 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:36:39 crc kubenswrapper[4894]: E1209 16:36:39.107819 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:36:52 crc kubenswrapper[4894]: I1209 16:36:52.106193 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:36:52 crc kubenswrapper[4894]: E1209 16:36:52.107028 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:37:03 crc kubenswrapper[4894]: I1209 16:37:03.106304 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:37:03 crc kubenswrapper[4894]: E1209 16:37:03.107258 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:37:17 crc kubenswrapper[4894]: I1209 16:37:17.106799 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:37:18 crc kubenswrapper[4894]: I1209 16:37:18.000734 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerStarted","Data":"09a30e3fcdd74db8135435e7d3692a70be84fb1b91c0d6ae45b63f5970ba4ceb"} Dec 09 16:37:37 crc kubenswrapper[4894]: I1209 16:37:37.166695 4894 generic.go:334] "Generic (PLEG): container finished" podID="3dd4c51e-1b24-48e9-a722-80c11a2460fd" containerID="8572fc20ed4efd6d16c60ad7c98da04cb647ff26aac0271aa26134c9ca70fc6f" exitCode=0 Dec 09 16:37:37 crc kubenswrapper[4894]: I1209 16:37:37.166753 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"3dd4c51e-1b24-48e9-a722-80c11a2460fd","Type":"ContainerDied","Data":"8572fc20ed4efd6d16c60ad7c98da04cb647ff26aac0271aa26134c9ca70fc6f"} Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.702683 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.857139 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-test-operator-clouds-config\") pod \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.857200 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.857255 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-tobiko-private-key\") pod \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.857307 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3dd4c51e-1b24-48e9-a722-80c11a2460fd-test-operator-ephemeral-temporary\") pod \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.857358 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-tobiko-public-key\") pod \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.857388 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-ca-certs\") pod \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.857416 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-tobiko-config\") pod \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.857531 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-openstack-config-secret\") pod \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.857572 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f66nv\" (UniqueName: \"kubernetes.io/projected/3dd4c51e-1b24-48e9-a722-80c11a2460fd-kube-api-access-f66nv\") pod \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.857596 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-ceph\") pod \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.857672 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-kubeconfig\") pod \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.857698 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3dd4c51e-1b24-48e9-a722-80c11a2460fd-test-operator-ephemeral-workdir\") pod \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\" (UID: \"3dd4c51e-1b24-48e9-a722-80c11a2460fd\") " Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.858321 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dd4c51e-1b24-48e9-a722-80c11a2460fd-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "3dd4c51e-1b24-48e9-a722-80c11a2460fd" (UID: "3dd4c51e-1b24-48e9-a722-80c11a2460fd"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.865624 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-ceph" (OuterVolumeSpecName: "ceph") pod "3dd4c51e-1b24-48e9-a722-80c11a2460fd" (UID: "3dd4c51e-1b24-48e9-a722-80c11a2460fd"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.865922 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3dd4c51e-1b24-48e9-a722-80c11a2460fd-kube-api-access-f66nv" (OuterVolumeSpecName: "kube-api-access-f66nv") pod "3dd4c51e-1b24-48e9-a722-80c11a2460fd" (UID: "3dd4c51e-1b24-48e9-a722-80c11a2460fd"). InnerVolumeSpecName "kube-api-access-f66nv". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.868986 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage01-crc" (OuterVolumeSpecName: "test-operator-logs") pod "3dd4c51e-1b24-48e9-a722-80c11a2460fd" (UID: "3dd4c51e-1b24-48e9-a722-80c11a2460fd"). InnerVolumeSpecName "local-storage01-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.889666 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "3dd4c51e-1b24-48e9-a722-80c11a2460fd" (UID: "3dd4c51e-1b24-48e9-a722-80c11a2460fd"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.890429 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-tobiko-config" (OuterVolumeSpecName: "tobiko-config") pod "3dd4c51e-1b24-48e9-a722-80c11a2460fd" (UID: "3dd4c51e-1b24-48e9-a722-80c11a2460fd"). InnerVolumeSpecName "tobiko-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.895481 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-tobiko-private-key" (OuterVolumeSpecName: "tobiko-private-key") pod "3dd4c51e-1b24-48e9-a722-80c11a2460fd" (UID: "3dd4c51e-1b24-48e9-a722-80c11a2460fd"). InnerVolumeSpecName "tobiko-private-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.900570 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-tobiko-public-key" (OuterVolumeSpecName: "tobiko-public-key") pod "3dd4c51e-1b24-48e9-a722-80c11a2460fd" (UID: "3dd4c51e-1b24-48e9-a722-80c11a2460fd"). InnerVolumeSpecName "tobiko-public-key". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.908856 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-kubeconfig" (OuterVolumeSpecName: "kubeconfig") pod "3dd4c51e-1b24-48e9-a722-80c11a2460fd" (UID: "3dd4c51e-1b24-48e9-a722-80c11a2460fd"). InnerVolumeSpecName "kubeconfig". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.920166 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-test-operator-clouds-config" (OuterVolumeSpecName: "test-operator-clouds-config") pod "3dd4c51e-1b24-48e9-a722-80c11a2460fd" (UID: "3dd4c51e-1b24-48e9-a722-80c11a2460fd"). InnerVolumeSpecName "test-operator-clouds-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.924629 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "3dd4c51e-1b24-48e9-a722-80c11a2460fd" (UID: "3dd4c51e-1b24-48e9-a722-80c11a2460fd"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.959870 4894 reconciler_common.go:293] "Volume detached for volume \"kubeconfig\" (UniqueName: \"kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-kubeconfig\") on node \"crc\" DevicePath \"\"" Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.959917 4894 reconciler_common.go:293] "Volume detached for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-test-operator-clouds-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.959946 4894 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" " Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.959955 4894 reconciler_common.go:293] "Volume detached for volume \"tobiko-private-key\" (UniqueName: \"kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-tobiko-private-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.959966 4894 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/3dd4c51e-1b24-48e9-a722-80c11a2460fd-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.959975 4894 reconciler_common.go:293] "Volume detached for volume \"tobiko-public-key\" (UniqueName: \"kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-tobiko-public-key\") on node \"crc\" DevicePath \"\"" Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.959984 4894 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.959993 4894 reconciler_common.go:293] "Volume detached for volume \"tobiko-config\" (UniqueName: \"kubernetes.io/configmap/3dd4c51e-1b24-48e9-a722-80c11a2460fd-tobiko-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.960003 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.960014 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f66nv\" (UniqueName: \"kubernetes.io/projected/3dd4c51e-1b24-48e9-a722-80c11a2460fd-kube-api-access-f66nv\") on node \"crc\" DevicePath \"\"" Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.960025 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/3dd4c51e-1b24-48e9-a722-80c11a2460fd-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:37:38 crc kubenswrapper[4894]: I1209 16:37:38.984680 4894 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage01-crc" (UniqueName: "kubernetes.io/local-volume/local-storage01-crc") on node "crc" Dec 09 16:37:39 crc kubenswrapper[4894]: I1209 16:37:39.062126 4894 reconciler_common.go:293] "Volume detached for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") on node \"crc\" DevicePath \"\"" Dec 09 16:37:39 crc kubenswrapper[4894]: I1209 16:37:39.185980 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/tobiko-tests-tobiko-s01-sanity" event={"ID":"3dd4c51e-1b24-48e9-a722-80c11a2460fd","Type":"ContainerDied","Data":"32ae6c77da3557f1ed549340b7f0c52b5e67b25b21fb09a0089013eeeb8e89fe"} Dec 09 16:37:39 crc kubenswrapper[4894]: I1209 16:37:39.186025 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="32ae6c77da3557f1ed549340b7f0c52b5e67b25b21fb09a0089013eeeb8e89fe" Dec 09 16:37:39 crc kubenswrapper[4894]: I1209 16:37:39.186035 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/tobiko-tests-tobiko-s01-sanity" Dec 09 16:37:40 crc kubenswrapper[4894]: I1209 16:37:40.196448 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3dd4c51e-1b24-48e9-a722-80c11a2460fd-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "3dd4c51e-1b24-48e9-a722-80c11a2460fd" (UID: "3dd4c51e-1b24-48e9-a722-80c11a2460fd"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:37:40 crc kubenswrapper[4894]: I1209 16:37:40.291391 4894 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/3dd4c51e-1b24-48e9-a722-80c11a2460fd-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 09 16:37:50 crc kubenswrapper[4894]: I1209 16:37:50.945735 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko"] Dec 09 16:37:50 crc kubenswrapper[4894]: E1209 16:37:50.946804 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1140abe-2fbd-449f-be06-e6581f5757c2" containerName="extract-utilities" Dec 09 16:37:50 crc kubenswrapper[4894]: I1209 16:37:50.946818 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1140abe-2fbd-449f-be06-e6581f5757c2" containerName="extract-utilities" Dec 09 16:37:50 crc kubenswrapper[4894]: E1209 16:37:50.946841 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1140abe-2fbd-449f-be06-e6581f5757c2" containerName="extract-content" Dec 09 16:37:50 crc kubenswrapper[4894]: I1209 16:37:50.946849 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1140abe-2fbd-449f-be06-e6581f5757c2" containerName="extract-content" Dec 09 16:37:50 crc kubenswrapper[4894]: E1209 16:37:50.946862 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3dd4c51e-1b24-48e9-a722-80c11a2460fd" containerName="tobiko-tests-tobiko" Dec 09 16:37:50 crc kubenswrapper[4894]: I1209 16:37:50.946869 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="3dd4c51e-1b24-48e9-a722-80c11a2460fd" containerName="tobiko-tests-tobiko" Dec 09 16:37:50 crc kubenswrapper[4894]: E1209 16:37:50.946885 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a1140abe-2fbd-449f-be06-e6581f5757c2" containerName="registry-server" Dec 09 16:37:50 crc kubenswrapper[4894]: I1209 16:37:50.946891 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="a1140abe-2fbd-449f-be06-e6581f5757c2" containerName="registry-server" Dec 09 16:37:50 crc kubenswrapper[4894]: I1209 16:37:50.947119 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="a1140abe-2fbd-449f-be06-e6581f5757c2" containerName="registry-server" Dec 09 16:37:50 crc kubenswrapper[4894]: I1209 16:37:50.947146 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="3dd4c51e-1b24-48e9-a722-80c11a2460fd" containerName="tobiko-tests-tobiko" Dec 09 16:37:50 crc kubenswrapper[4894]: I1209 16:37:50.948061 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Dec 09 16:37:50 crc kubenswrapper[4894]: I1209 16:37:50.960487 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko"] Dec 09 16:37:51 crc kubenswrapper[4894]: I1209 16:37:51.100808 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmj2x\" (UniqueName: \"kubernetes.io/projected/8135eb77-f9c3-4570-b4d8-c7d87f8cdf52-kube-api-access-cmj2x\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"8135eb77-f9c3-4570-b4d8-c7d87f8cdf52\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Dec 09 16:37:51 crc kubenswrapper[4894]: I1209 16:37:51.101216 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"8135eb77-f9c3-4570-b4d8-c7d87f8cdf52\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Dec 09 16:37:51 crc kubenswrapper[4894]: I1209 16:37:51.204287 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmj2x\" (UniqueName: \"kubernetes.io/projected/8135eb77-f9c3-4570-b4d8-c7d87f8cdf52-kube-api-access-cmj2x\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"8135eb77-f9c3-4570-b4d8-c7d87f8cdf52\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Dec 09 16:37:51 crc kubenswrapper[4894]: I1209 16:37:51.204371 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"8135eb77-f9c3-4570-b4d8-c7d87f8cdf52\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Dec 09 16:37:51 crc kubenswrapper[4894]: I1209 16:37:51.206688 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"8135eb77-f9c3-4570-b4d8-c7d87f8cdf52\") device mount path \"/mnt/openstack/pv01\"" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Dec 09 16:37:51 crc kubenswrapper[4894]: I1209 16:37:51.235296 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmj2x\" (UniqueName: \"kubernetes.io/projected/8135eb77-f9c3-4570-b4d8-c7d87f8cdf52-kube-api-access-cmj2x\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"8135eb77-f9c3-4570-b4d8-c7d87f8cdf52\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Dec 09 16:37:51 crc kubenswrapper[4894]: I1209 16:37:51.239553 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"test-operator-logs-pod-tobiko-tobiko-tests-tobiko\" (UID: \"8135eb77-f9c3-4570-b4d8-c7d87f8cdf52\") " pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Dec 09 16:37:51 crc kubenswrapper[4894]: I1209 16:37:51.270372 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" Dec 09 16:37:51 crc kubenswrapper[4894]: I1209 16:37:51.693163 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko"] Dec 09 16:37:51 crc kubenswrapper[4894]: I1209 16:37:51.696676 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 16:37:52 crc kubenswrapper[4894]: I1209 16:37:52.311269 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" event={"ID":"8135eb77-f9c3-4570-b4d8-c7d87f8cdf52","Type":"ContainerStarted","Data":"4d1f593cba71ebb980d71aba2a8443d098bec6947084f081525440cbc1ca040c"} Dec 09 16:37:53 crc kubenswrapper[4894]: I1209 16:37:53.323242 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" event={"ID":"8135eb77-f9c3-4570-b4d8-c7d87f8cdf52","Type":"ContainerStarted","Data":"273b54f84745be368b9caee3c1d3a036b4945632de5481a71f2688ee7ab37321"} Dec 09 16:37:53 crc kubenswrapper[4894]: I1209 16:37:53.344244 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-tobiko-tobiko-tests-tobiko" podStartSLOduration=2.551584561 podStartE2EDuration="3.344221313s" podCreationTimestamp="2025-12-09 16:37:50 +0000 UTC" firstStartedPulling="2025-12-09 16:37:51.696474009 +0000 UTC m=+3966.015684668" lastFinishedPulling="2025-12-09 16:37:52.489110741 +0000 UTC m=+3966.808321420" observedRunningTime="2025-12-09 16:37:53.336158224 +0000 UTC m=+3967.655368913" watchObservedRunningTime="2025-12-09 16:37:53.344221313 +0000 UTC m=+3967.663431992" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.493552 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ansibletest-ansibletest"] Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.495861 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.505511 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"test-operator-controller-priv-key" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.505932 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dataplane-ansible-ssh-private-key-secret" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.514936 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ansibletest-ansibletest"] Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.588459 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-compute-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.588630 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kkmf\" (UniqueName: \"kubernetes.io/projected/b56c2145-351c-42cf-8a4e-049c7825d527-kube-api-access-7kkmf\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.588797 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-openstack-config-secret\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.588906 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.589091 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b56c2145-351c-42cf-8a4e-049c7825d527-test-operator-ephemeral-temporary\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.589140 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-workload-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.589180 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b56c2145-351c-42cf-8a4e-049c7825d527-test-operator-ephemeral-workdir\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.589466 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b56c2145-351c-42cf-8a4e-049c7825d527-openstack-config\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.589669 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-ca-certs\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.589767 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-ceph\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.692105 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-ceph\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.692364 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-compute-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.692486 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kkmf\" (UniqueName: \"kubernetes.io/projected/b56c2145-351c-42cf-8a4e-049c7825d527-kube-api-access-7kkmf\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.692564 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-openstack-config-secret\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.692628 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.692801 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b56c2145-351c-42cf-8a4e-049c7825d527-test-operator-ephemeral-temporary\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.692865 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-workload-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.692913 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b56c2145-351c-42cf-8a4e-049c7825d527-test-operator-ephemeral-workdir\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.693034 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b56c2145-351c-42cf-8a4e-049c7825d527-openstack-config\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.693122 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-ca-certs\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.694067 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.694078 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b56c2145-351c-42cf-8a4e-049c7825d527-test-operator-ephemeral-temporary\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.694333 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b56c2145-351c-42cf-8a4e-049c7825d527-test-operator-ephemeral-workdir\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.695076 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b56c2145-351c-42cf-8a4e-049c7825d527-openstack-config\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.701669 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-ceph\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.702147 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-openstack-config-secret\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.705161 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-ca-certs\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.705830 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-workload-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.706099 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-compute-ssh-secret\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.714492 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kkmf\" (UniqueName: \"kubernetes.io/projected/b56c2145-351c-42cf-8a4e-049c7825d527-kube-api-access-7kkmf\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.736913 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"ansibletest-ansibletest\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " pod="openstack/ansibletest-ansibletest" Dec 09 16:38:05 crc kubenswrapper[4894]: I1209 16:38:05.851311 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Dec 09 16:38:06 crc kubenswrapper[4894]: I1209 16:38:06.379834 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ansibletest-ansibletest"] Dec 09 16:38:06 crc kubenswrapper[4894]: I1209 16:38:06.460501 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"b56c2145-351c-42cf-8a4e-049c7825d527","Type":"ContainerStarted","Data":"4079c1ed0e003b1666df8b3c1c9c9fa50631e453fa05ea16bcb4d08dd175fdd7"} Dec 09 16:38:43 crc kubenswrapper[4894]: E1209 16:38:43.087613 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-ansible-tests:current-podified" Dec 09 16:38:43 crc kubenswrapper[4894]: E1209 16:38:43.088352 4894 kuberuntime_manager.go:1274] "Unhandled Error" err=< Dec 09 16:38:43 crc kubenswrapper[4894]: container &Container{Name:ansibletest-ansibletest,Image:quay.io/podified-antelope-centos9/openstack-ansible-tests:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:POD_ANSIBLE_EXTRA_VARS,Value:-e manual_run=false,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_FILE_EXTRA_VARS,Value:--- Dec 09 16:38:43 crc kubenswrapper[4894]: foo: bar Dec 09 16:38:43 crc kubenswrapper[4894]: ,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_GIT_BRANCH,Value:,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_GIT_REPO,Value:https://github.com/ansible/test-playbooks,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_INVENTORY,Value:localhost ansible_connection=local ansible_python_interpreter=python3 Dec 09 16:38:43 crc kubenswrapper[4894]: ,ValueFrom:nil,},EnvVar{Name:POD_ANSIBLE_PLAYBOOK,Value:./debug.yml,ValueFrom:nil,},EnvVar{Name:POD_DEBUG,Value:false,ValueFrom:nil,},EnvVar{Name:POD_INSTALL_COLLECTIONS,Value:,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{4 0} {} 4 DecimalSI},memory: {{4294967296 0} {} 4Gi BinarySI},},Requests:ResourceList{cpu: {{2 0} {} 2 DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/ansible,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/AnsibleTests/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config,ReadOnly:true,MountPath:/var/lib/ansible/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/var/lib/ansible/.config/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ca-bundle.trust.crt,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:compute-ssh-secret,ReadOnly:true,MountPath:/var/lib/ansible/.ssh/compute_id,SubPath:ssh-privatekey,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:workload-ssh-secret,ReadOnly:true,MountPath:/var/lib/ansible/test_keypair.key,SubPath:ssh-privatekey,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-7kkmf,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN NET_RAW],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*227,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*227,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod ansibletest-ansibletest_openstack(b56c2145-351c-42cf-8a4e-049c7825d527): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled Dec 09 16:38:43 crc kubenswrapper[4894]: > logger="UnhandledError" Dec 09 16:38:43 crc kubenswrapper[4894]: E1209 16:38:43.090090 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ansibletest-ansibletest\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/ansibletest-ansibletest" podUID="b56c2145-351c-42cf-8a4e-049c7825d527" Dec 09 16:38:43 crc kubenswrapper[4894]: E1209 16:38:43.881241 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ansibletest-ansibletest\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-ansible-tests:current-podified\\\"\"" pod="openstack/ansibletest-ansibletest" podUID="b56c2145-351c-42cf-8a4e-049c7825d527" Dec 09 16:38:58 crc kubenswrapper[4894]: I1209 16:38:58.032879 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"b56c2145-351c-42cf-8a4e-049c7825d527","Type":"ContainerStarted","Data":"26e68345b38d71b7fec27f25b0ddecbd746dae606c4a2255498a38717cb7fb03"} Dec 09 16:38:58 crc kubenswrapper[4894]: I1209 16:38:58.052536 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ansibletest-ansibletest" podStartSLOduration=3.291709633 podStartE2EDuration="54.052519908s" podCreationTimestamp="2025-12-09 16:38:04 +0000 UTC" firstStartedPulling="2025-12-09 16:38:06.383382593 +0000 UTC m=+3980.702593272" lastFinishedPulling="2025-12-09 16:38:57.144192838 +0000 UTC m=+4031.463403547" observedRunningTime="2025-12-09 16:38:58.049055534 +0000 UTC m=+4032.368266223" watchObservedRunningTime="2025-12-09 16:38:58.052519908 +0000 UTC m=+4032.371730577" Dec 09 16:39:00 crc kubenswrapper[4894]: I1209 16:39:00.054062 4894 generic.go:334] "Generic (PLEG): container finished" podID="b56c2145-351c-42cf-8a4e-049c7825d527" containerID="26e68345b38d71b7fec27f25b0ddecbd746dae606c4a2255498a38717cb7fb03" exitCode=0 Dec 09 16:39:00 crc kubenswrapper[4894]: I1209 16:39:00.054165 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"b56c2145-351c-42cf-8a4e-049c7825d527","Type":"ContainerDied","Data":"26e68345b38d71b7fec27f25b0ddecbd746dae606c4a2255498a38717cb7fb03"} Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.663435 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.825198 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-ca-certs\") pod \"b56c2145-351c-42cf-8a4e-049c7825d527\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.825260 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-ceph\") pod \"b56c2145-351c-42cf-8a4e-049c7825d527\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.825291 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-compute-ssh-secret\") pod \"b56c2145-351c-42cf-8a4e-049c7825d527\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.825331 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7kkmf\" (UniqueName: \"kubernetes.io/projected/b56c2145-351c-42cf-8a4e-049c7825d527-kube-api-access-7kkmf\") pod \"b56c2145-351c-42cf-8a4e-049c7825d527\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.825367 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-workload-ssh-secret\") pod \"b56c2145-351c-42cf-8a4e-049c7825d527\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.825419 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b56c2145-351c-42cf-8a4e-049c7825d527-test-operator-ephemeral-workdir\") pod \"b56c2145-351c-42cf-8a4e-049c7825d527\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.825442 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b56c2145-351c-42cf-8a4e-049c7825d527-test-operator-ephemeral-temporary\") pod \"b56c2145-351c-42cf-8a4e-049c7825d527\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.825502 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"b56c2145-351c-42cf-8a4e-049c7825d527\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.825534 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b56c2145-351c-42cf-8a4e-049c7825d527-openstack-config\") pod \"b56c2145-351c-42cf-8a4e-049c7825d527\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.825567 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-openstack-config-secret\") pod \"b56c2145-351c-42cf-8a4e-049c7825d527\" (UID: \"b56c2145-351c-42cf-8a4e-049c7825d527\") " Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.827149 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b56c2145-351c-42cf-8a4e-049c7825d527-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "b56c2145-351c-42cf-8a4e-049c7825d527" (UID: "b56c2145-351c-42cf-8a4e-049c7825d527"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.835821 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-ceph" (OuterVolumeSpecName: "ceph") pod "b56c2145-351c-42cf-8a4e-049c7825d527" (UID: "b56c2145-351c-42cf-8a4e-049c7825d527"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.835833 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "test-operator-logs") pod "b56c2145-351c-42cf-8a4e-049c7825d527" (UID: "b56c2145-351c-42cf-8a4e-049c7825d527"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.835829 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b56c2145-351c-42cf-8a4e-049c7825d527-kube-api-access-7kkmf" (OuterVolumeSpecName: "kube-api-access-7kkmf") pod "b56c2145-351c-42cf-8a4e-049c7825d527" (UID: "b56c2145-351c-42cf-8a4e-049c7825d527"). InnerVolumeSpecName "kube-api-access-7kkmf". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.842441 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b56c2145-351c-42cf-8a4e-049c7825d527-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "b56c2145-351c-42cf-8a4e-049c7825d527" (UID: "b56c2145-351c-42cf-8a4e-049c7825d527"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.862626 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "b56c2145-351c-42cf-8a4e-049c7825d527" (UID: "b56c2145-351c-42cf-8a4e-049c7825d527"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.879337 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-compute-ssh-secret" (OuterVolumeSpecName: "compute-ssh-secret") pod "b56c2145-351c-42cf-8a4e-049c7825d527" (UID: "b56c2145-351c-42cf-8a4e-049c7825d527"). InnerVolumeSpecName "compute-ssh-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.881514 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b56c2145-351c-42cf-8a4e-049c7825d527-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "b56c2145-351c-42cf-8a4e-049c7825d527" (UID: "b56c2145-351c-42cf-8a4e-049c7825d527"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.893402 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-workload-ssh-secret" (OuterVolumeSpecName: "workload-ssh-secret") pod "b56c2145-351c-42cf-8a4e-049c7825d527" (UID: "b56c2145-351c-42cf-8a4e-049c7825d527"). InnerVolumeSpecName "workload-ssh-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.893717 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "b56c2145-351c-42cf-8a4e-049c7825d527" (UID: "b56c2145-351c-42cf-8a4e-049c7825d527"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.928050 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.928142 4894 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.928183 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.928243 4894 reconciler_common.go:293] "Volume detached for volume \"compute-ssh-secret\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-compute-ssh-secret\") on node \"crc\" DevicePath \"\"" Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.928256 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7kkmf\" (UniqueName: \"kubernetes.io/projected/b56c2145-351c-42cf-8a4e-049c7825d527-kube-api-access-7kkmf\") on node \"crc\" DevicePath \"\"" Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.928268 4894 reconciler_common.go:293] "Volume detached for volume \"workload-ssh-secret\" (UniqueName: \"kubernetes.io/secret/b56c2145-351c-42cf-8a4e-049c7825d527-workload-ssh-secret\") on node \"crc\" DevicePath \"\"" Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.928306 4894 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/b56c2145-351c-42cf-8a4e-049c7825d527-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.928321 4894 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/b56c2145-351c-42cf-8a4e-049c7825d527-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.928376 4894 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.928395 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/b56c2145-351c-42cf-8a4e-049c7825d527-openstack-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:39:02 crc kubenswrapper[4894]: I1209 16:39:02.956773 4894 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Dec 09 16:39:03 crc kubenswrapper[4894]: I1209 16:39:03.030282 4894 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Dec 09 16:39:03 crc kubenswrapper[4894]: I1209 16:39:03.083580 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ansibletest-ansibletest" event={"ID":"b56c2145-351c-42cf-8a4e-049c7825d527","Type":"ContainerDied","Data":"4079c1ed0e003b1666df8b3c1c9c9fa50631e453fa05ea16bcb4d08dd175fdd7"} Dec 09 16:39:03 crc kubenswrapper[4894]: I1209 16:39:03.083621 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4079c1ed0e003b1666df8b3c1c9c9fa50631e453fa05ea16bcb4d08dd175fdd7" Dec 09 16:39:03 crc kubenswrapper[4894]: I1209 16:39:03.083688 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ansibletest-ansibletest" Dec 09 16:39:05 crc kubenswrapper[4894]: I1209 16:39:05.159959 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-88kvs"] Dec 09 16:39:05 crc kubenswrapper[4894]: E1209 16:39:05.161010 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b56c2145-351c-42cf-8a4e-049c7825d527" containerName="ansibletest-ansibletest" Dec 09 16:39:05 crc kubenswrapper[4894]: I1209 16:39:05.161029 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b56c2145-351c-42cf-8a4e-049c7825d527" containerName="ansibletest-ansibletest" Dec 09 16:39:05 crc kubenswrapper[4894]: I1209 16:39:05.161292 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b56c2145-351c-42cf-8a4e-049c7825d527" containerName="ansibletest-ansibletest" Dec 09 16:39:05 crc kubenswrapper[4894]: I1209 16:39:05.163139 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-88kvs" Dec 09 16:39:05 crc kubenswrapper[4894]: I1209 16:39:05.196578 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-88kvs"] Dec 09 16:39:05 crc kubenswrapper[4894]: I1209 16:39:05.275276 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da62df93-33dd-4f5c-85f6-e82f8b1f1b2a-utilities\") pod \"redhat-operators-88kvs\" (UID: \"da62df93-33dd-4f5c-85f6-e82f8b1f1b2a\") " pod="openshift-marketplace/redhat-operators-88kvs" Dec 09 16:39:05 crc kubenswrapper[4894]: I1209 16:39:05.275470 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl8xg\" (UniqueName: \"kubernetes.io/projected/da62df93-33dd-4f5c-85f6-e82f8b1f1b2a-kube-api-access-gl8xg\") pod \"redhat-operators-88kvs\" (UID: \"da62df93-33dd-4f5c-85f6-e82f8b1f1b2a\") " pod="openshift-marketplace/redhat-operators-88kvs" Dec 09 16:39:05 crc kubenswrapper[4894]: I1209 16:39:05.275582 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da62df93-33dd-4f5c-85f6-e82f8b1f1b2a-catalog-content\") pod \"redhat-operators-88kvs\" (UID: \"da62df93-33dd-4f5c-85f6-e82f8b1f1b2a\") " pod="openshift-marketplace/redhat-operators-88kvs" Dec 09 16:39:05 crc kubenswrapper[4894]: I1209 16:39:05.378125 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da62df93-33dd-4f5c-85f6-e82f8b1f1b2a-utilities\") pod \"redhat-operators-88kvs\" (UID: \"da62df93-33dd-4f5c-85f6-e82f8b1f1b2a\") " pod="openshift-marketplace/redhat-operators-88kvs" Dec 09 16:39:05 crc kubenswrapper[4894]: I1209 16:39:05.378266 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl8xg\" (UniqueName: \"kubernetes.io/projected/da62df93-33dd-4f5c-85f6-e82f8b1f1b2a-kube-api-access-gl8xg\") pod \"redhat-operators-88kvs\" (UID: \"da62df93-33dd-4f5c-85f6-e82f8b1f1b2a\") " pod="openshift-marketplace/redhat-operators-88kvs" Dec 09 16:39:05 crc kubenswrapper[4894]: I1209 16:39:05.378342 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da62df93-33dd-4f5c-85f6-e82f8b1f1b2a-catalog-content\") pod \"redhat-operators-88kvs\" (UID: \"da62df93-33dd-4f5c-85f6-e82f8b1f1b2a\") " pod="openshift-marketplace/redhat-operators-88kvs" Dec 09 16:39:05 crc kubenswrapper[4894]: I1209 16:39:05.379039 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da62df93-33dd-4f5c-85f6-e82f8b1f1b2a-catalog-content\") pod \"redhat-operators-88kvs\" (UID: \"da62df93-33dd-4f5c-85f6-e82f8b1f1b2a\") " pod="openshift-marketplace/redhat-operators-88kvs" Dec 09 16:39:05 crc kubenswrapper[4894]: I1209 16:39:05.379424 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da62df93-33dd-4f5c-85f6-e82f8b1f1b2a-utilities\") pod \"redhat-operators-88kvs\" (UID: \"da62df93-33dd-4f5c-85f6-e82f8b1f1b2a\") " pod="openshift-marketplace/redhat-operators-88kvs" Dec 09 16:39:05 crc kubenswrapper[4894]: I1209 16:39:05.412476 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl8xg\" (UniqueName: \"kubernetes.io/projected/da62df93-33dd-4f5c-85f6-e82f8b1f1b2a-kube-api-access-gl8xg\") pod \"redhat-operators-88kvs\" (UID: \"da62df93-33dd-4f5c-85f6-e82f8b1f1b2a\") " pod="openshift-marketplace/redhat-operators-88kvs" Dec 09 16:39:05 crc kubenswrapper[4894]: I1209 16:39:05.491716 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-88kvs" Dec 09 16:39:05 crc kubenswrapper[4894]: I1209 16:39:05.959038 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-88kvs"] Dec 09 16:39:06 crc kubenswrapper[4894]: I1209 16:39:06.120566 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-88kvs" event={"ID":"da62df93-33dd-4f5c-85f6-e82f8b1f1b2a","Type":"ContainerStarted","Data":"6aeac3d18a3cebe98ccb1b21ad89583af55de6d9c29c4bc93a1cf3eddbfab58d"} Dec 09 16:39:07 crc kubenswrapper[4894]: I1209 16:39:07.131856 4894 generic.go:334] "Generic (PLEG): container finished" podID="da62df93-33dd-4f5c-85f6-e82f8b1f1b2a" containerID="f2789b81f06f58c1b06486ee90fb210e5e8a3c59e3cca5ec013650970b9409e0" exitCode=0 Dec 09 16:39:07 crc kubenswrapper[4894]: I1209 16:39:07.131933 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-88kvs" event={"ID":"da62df93-33dd-4f5c-85f6-e82f8b1f1b2a","Type":"ContainerDied","Data":"f2789b81f06f58c1b06486ee90fb210e5e8a3c59e3cca5ec013650970b9409e0"} Dec 09 16:39:11 crc kubenswrapper[4894]: I1209 16:39:11.936133 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest"] Dec 09 16:39:11 crc kubenswrapper[4894]: I1209 16:39:11.942453 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Dec 09 16:39:11 crc kubenswrapper[4894]: I1209 16:39:11.961148 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest"] Dec 09 16:39:12 crc kubenswrapper[4894]: I1209 16:39:12.047924 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgpjv\" (UniqueName: \"kubernetes.io/projected/b81b27e0-9170-4753-97ec-5354f447d801-kube-api-access-dgpjv\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"b81b27e0-9170-4753-97ec-5354f447d801\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Dec 09 16:39:12 crc kubenswrapper[4894]: I1209 16:39:12.048103 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"b81b27e0-9170-4753-97ec-5354f447d801\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Dec 09 16:39:12 crc kubenswrapper[4894]: I1209 16:39:12.149930 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"b81b27e0-9170-4753-97ec-5354f447d801\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Dec 09 16:39:12 crc kubenswrapper[4894]: I1209 16:39:12.150075 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgpjv\" (UniqueName: \"kubernetes.io/projected/b81b27e0-9170-4753-97ec-5354f447d801-kube-api-access-dgpjv\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"b81b27e0-9170-4753-97ec-5354f447d801\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Dec 09 16:39:12 crc kubenswrapper[4894]: I1209 16:39:12.150607 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"b81b27e0-9170-4753-97ec-5354f447d801\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Dec 09 16:39:12 crc kubenswrapper[4894]: I1209 16:39:12.188949 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgpjv\" (UniqueName: \"kubernetes.io/projected/b81b27e0-9170-4753-97ec-5354f447d801-kube-api-access-dgpjv\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"b81b27e0-9170-4753-97ec-5354f447d801\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Dec 09 16:39:12 crc kubenswrapper[4894]: I1209 16:39:12.200438 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"test-operator-logs-pod-ansibletest-ansibletest-ansibletest\" (UID: \"b81b27e0-9170-4753-97ec-5354f447d801\") " pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Dec 09 16:39:12 crc kubenswrapper[4894]: I1209 16:39:12.262044 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" Dec 09 16:39:18 crc kubenswrapper[4894]: I1209 16:39:18.209930 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest"] Dec 09 16:39:18 crc kubenswrapper[4894]: I1209 16:39:18.264248 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" event={"ID":"b81b27e0-9170-4753-97ec-5354f447d801","Type":"ContainerStarted","Data":"b9f592db142d43954cc5a51026c7193e16cfff01ddced45250f0a49a549cfa9d"} Dec 09 16:39:18 crc kubenswrapper[4894]: I1209 16:39:18.267160 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-88kvs" event={"ID":"da62df93-33dd-4f5c-85f6-e82f8b1f1b2a","Type":"ContainerStarted","Data":"9c8812f487c221b84c5063e8e7a7323e98715865c479e6da9f67f9fd6d5e0a10"} Dec 09 16:39:21 crc kubenswrapper[4894]: I1209 16:39:21.302324 4894 generic.go:334] "Generic (PLEG): container finished" podID="da62df93-33dd-4f5c-85f6-e82f8b1f1b2a" containerID="9c8812f487c221b84c5063e8e7a7323e98715865c479e6da9f67f9fd6d5e0a10" exitCode=0 Dec 09 16:39:21 crc kubenswrapper[4894]: I1209 16:39:21.302397 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-88kvs" event={"ID":"da62df93-33dd-4f5c-85f6-e82f8b1f1b2a","Type":"ContainerDied","Data":"9c8812f487c221b84c5063e8e7a7323e98715865c479e6da9f67f9fd6d5e0a10"} Dec 09 16:39:23 crc kubenswrapper[4894]: I1209 16:39:23.325172 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" event={"ID":"b81b27e0-9170-4753-97ec-5354f447d801","Type":"ContainerStarted","Data":"fbb1bef36855c540977d25190f5c7427f87f608d213e3f6274b0693acb4571ad"} Dec 09 16:39:23 crc kubenswrapper[4894]: I1209 16:39:23.364463 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-ansibletest-ansibletest-ansibletest" podStartSLOduration=8.552627827 podStartE2EDuration="12.364442519s" podCreationTimestamp="2025-12-09 16:39:11 +0000 UTC" firstStartedPulling="2025-12-09 16:39:18.22257991 +0000 UTC m=+4052.541790579" lastFinishedPulling="2025-12-09 16:39:22.034394592 +0000 UTC m=+4056.353605271" observedRunningTime="2025-12-09 16:39:23.339370918 +0000 UTC m=+4057.658581607" watchObservedRunningTime="2025-12-09 16:39:23.364442519 +0000 UTC m=+4057.683653178" Dec 09 16:39:26 crc kubenswrapper[4894]: I1209 16:39:26.357393 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-88kvs" event={"ID":"da62df93-33dd-4f5c-85f6-e82f8b1f1b2a","Type":"ContainerStarted","Data":"8d15acaa9fa044376c85b6f2276493280a74117a8d13a45908613a6af3ff4dea"} Dec 09 16:39:26 crc kubenswrapper[4894]: I1209 16:39:26.391179 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-88kvs" podStartSLOduration=6.002467699 podStartE2EDuration="21.391151444s" podCreationTimestamp="2025-12-09 16:39:05 +0000 UTC" firstStartedPulling="2025-12-09 16:39:07.134913961 +0000 UTC m=+4041.454124630" lastFinishedPulling="2025-12-09 16:39:22.523597706 +0000 UTC m=+4056.842808375" observedRunningTime="2025-12-09 16:39:26.382868758 +0000 UTC m=+4060.702079457" watchObservedRunningTime="2025-12-09 16:39:26.391151444 +0000 UTC m=+4060.710362123" Dec 09 16:39:35 crc kubenswrapper[4894]: I1209 16:39:35.492458 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-88kvs" Dec 09 16:39:35 crc kubenswrapper[4894]: I1209 16:39:35.493240 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-88kvs" Dec 09 16:39:35 crc kubenswrapper[4894]: I1209 16:39:35.587462 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-88kvs" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.513716 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-88kvs" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.597931 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-88kvs"] Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.613227 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/horizontest-tests-horizontest"] Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.616859 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.631333 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"test-operator-clouds-config" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.631397 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"horizontest-tests-horizontesthorizontest-config" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.666849 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizontest-tests-horizontest"] Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.696740 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jhjxg"] Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.697363 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-jhjxg" podUID="2a3d33af-2799-4c66-b963-3d6c70036854" containerName="registry-server" containerID="cri-o://f9efed26ba05df2e8a855d216e6a2aa579d27b9c00cedb9a568d118b94aee309" gracePeriod=2 Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.732066 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-test-operator-ephemeral-temporary\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.732170 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-openstack-config-secret\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.732249 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-test-operator-ephemeral-workdir\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.732310 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-ceph\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.732335 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-ca-certs\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.732482 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-test-operator-clouds-config\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.732524 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.732543 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qc64h\" (UniqueName: \"kubernetes.io/projected/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-kube-api-access-qc64h\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.833825 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-test-operator-clouds-config\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.834198 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.834218 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qc64h\" (UniqueName: \"kubernetes.io/projected/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-kube-api-access-qc64h\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.834248 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-test-operator-ephemeral-temporary\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.834288 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-openstack-config-secret\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.834344 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-test-operator-ephemeral-workdir\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.834383 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-ceph\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.834404 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-ca-certs\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.835049 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-test-operator-ephemeral-temporary\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.835215 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-test-operator-clouds-config\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.835326 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.835919 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-test-operator-ephemeral-workdir\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.841548 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-ca-certs\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.845107 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-openstack-config-secret\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.845128 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-ceph\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.852236 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qc64h\" (UniqueName: \"kubernetes.io/projected/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-kube-api-access-qc64h\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.905193 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"horizontest-tests-horizontest\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:36 crc kubenswrapper[4894]: I1209 16:39:36.969107 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Dec 09 16:39:37 crc kubenswrapper[4894]: I1209 16:39:37.450431 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/horizontest-tests-horizontest"] Dec 09 16:39:37 crc kubenswrapper[4894]: W1209 16:39:37.454467 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6573694d_aa1d_47bc_bfcd_6d92f7416f1a.slice/crio-9e4a8e993e89b3d75990dbe9ceadf9c0516bd528b359cbe476bcadc3569c819b WatchSource:0}: Error finding container 9e4a8e993e89b3d75990dbe9ceadf9c0516bd528b359cbe476bcadc3569c819b: Status 404 returned error can't find the container with id 9e4a8e993e89b3d75990dbe9ceadf9c0516bd528b359cbe476bcadc3569c819b Dec 09 16:39:37 crc kubenswrapper[4894]: I1209 16:39:37.476627 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"6573694d-aa1d-47bc-bfcd-6d92f7416f1a","Type":"ContainerStarted","Data":"9e4a8e993e89b3d75990dbe9ceadf9c0516bd528b359cbe476bcadc3569c819b"} Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.247167 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jhjxg" Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.377343 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xdmd2\" (UniqueName: \"kubernetes.io/projected/2a3d33af-2799-4c66-b963-3d6c70036854-kube-api-access-xdmd2\") pod \"2a3d33af-2799-4c66-b963-3d6c70036854\" (UID: \"2a3d33af-2799-4c66-b963-3d6c70036854\") " Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.377530 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a3d33af-2799-4c66-b963-3d6c70036854-utilities\") pod \"2a3d33af-2799-4c66-b963-3d6c70036854\" (UID: \"2a3d33af-2799-4c66-b963-3d6c70036854\") " Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.377650 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a3d33af-2799-4c66-b963-3d6c70036854-catalog-content\") pod \"2a3d33af-2799-4c66-b963-3d6c70036854\" (UID: \"2a3d33af-2799-4c66-b963-3d6c70036854\") " Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.379298 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a3d33af-2799-4c66-b963-3d6c70036854-utilities" (OuterVolumeSpecName: "utilities") pod "2a3d33af-2799-4c66-b963-3d6c70036854" (UID: "2a3d33af-2799-4c66-b963-3d6c70036854"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.386218 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a3d33af-2799-4c66-b963-3d6c70036854-kube-api-access-xdmd2" (OuterVolumeSpecName: "kube-api-access-xdmd2") pod "2a3d33af-2799-4c66-b963-3d6c70036854" (UID: "2a3d33af-2799-4c66-b963-3d6c70036854"). InnerVolumeSpecName "kube-api-access-xdmd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.479996 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xdmd2\" (UniqueName: \"kubernetes.io/projected/2a3d33af-2799-4c66-b963-3d6c70036854-kube-api-access-xdmd2\") on node \"crc\" DevicePath \"\"" Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.480033 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2a3d33af-2799-4c66-b963-3d6c70036854-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.490708 4894 generic.go:334] "Generic (PLEG): container finished" podID="2a3d33af-2799-4c66-b963-3d6c70036854" containerID="f9efed26ba05df2e8a855d216e6a2aa579d27b9c00cedb9a568d118b94aee309" exitCode=0 Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.491943 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhjxg" event={"ID":"2a3d33af-2799-4c66-b963-3d6c70036854","Type":"ContainerDied","Data":"f9efed26ba05df2e8a855d216e6a2aa579d27b9c00cedb9a568d118b94aee309"} Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.491982 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-jhjxg" event={"ID":"2a3d33af-2799-4c66-b963-3d6c70036854","Type":"ContainerDied","Data":"c5b042877a52da91891ce8e43cc987fa04ce95c8c06441d386033865cd594216"} Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.492035 4894 scope.go:117] "RemoveContainer" containerID="f9efed26ba05df2e8a855d216e6a2aa579d27b9c00cedb9a568d118b94aee309" Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.493091 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-jhjxg" Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.521774 4894 scope.go:117] "RemoveContainer" containerID="b8608fd119dd49bd3b5167b25d661a2c0a801cad26384f2f3c165a0ab4dbce88" Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.532136 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2a3d33af-2799-4c66-b963-3d6c70036854-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2a3d33af-2799-4c66-b963-3d6c70036854" (UID: "2a3d33af-2799-4c66-b963-3d6c70036854"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.566901 4894 scope.go:117] "RemoveContainer" containerID="0bb033a956a7bc057898f56b56faf95ed01c561ce6d3f07138ee162824bb5def" Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.614164 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2a3d33af-2799-4c66-b963-3d6c70036854-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.643819 4894 scope.go:117] "RemoveContainer" containerID="f9efed26ba05df2e8a855d216e6a2aa579d27b9c00cedb9a568d118b94aee309" Dec 09 16:39:38 crc kubenswrapper[4894]: E1209 16:39:38.645051 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f9efed26ba05df2e8a855d216e6a2aa579d27b9c00cedb9a568d118b94aee309\": container with ID starting with f9efed26ba05df2e8a855d216e6a2aa579d27b9c00cedb9a568d118b94aee309 not found: ID does not exist" containerID="f9efed26ba05df2e8a855d216e6a2aa579d27b9c00cedb9a568d118b94aee309" Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.645120 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f9efed26ba05df2e8a855d216e6a2aa579d27b9c00cedb9a568d118b94aee309"} err="failed to get container status \"f9efed26ba05df2e8a855d216e6a2aa579d27b9c00cedb9a568d118b94aee309\": rpc error: code = NotFound desc = could not find container \"f9efed26ba05df2e8a855d216e6a2aa579d27b9c00cedb9a568d118b94aee309\": container with ID starting with f9efed26ba05df2e8a855d216e6a2aa579d27b9c00cedb9a568d118b94aee309 not found: ID does not exist" Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.645167 4894 scope.go:117] "RemoveContainer" containerID="b8608fd119dd49bd3b5167b25d661a2c0a801cad26384f2f3c165a0ab4dbce88" Dec 09 16:39:38 crc kubenswrapper[4894]: E1209 16:39:38.647053 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8608fd119dd49bd3b5167b25d661a2c0a801cad26384f2f3c165a0ab4dbce88\": container with ID starting with b8608fd119dd49bd3b5167b25d661a2c0a801cad26384f2f3c165a0ab4dbce88 not found: ID does not exist" containerID="b8608fd119dd49bd3b5167b25d661a2c0a801cad26384f2f3c165a0ab4dbce88" Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.647105 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8608fd119dd49bd3b5167b25d661a2c0a801cad26384f2f3c165a0ab4dbce88"} err="failed to get container status \"b8608fd119dd49bd3b5167b25d661a2c0a801cad26384f2f3c165a0ab4dbce88\": rpc error: code = NotFound desc = could not find container \"b8608fd119dd49bd3b5167b25d661a2c0a801cad26384f2f3c165a0ab4dbce88\": container with ID starting with b8608fd119dd49bd3b5167b25d661a2c0a801cad26384f2f3c165a0ab4dbce88 not found: ID does not exist" Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.647136 4894 scope.go:117] "RemoveContainer" containerID="0bb033a956a7bc057898f56b56faf95ed01c561ce6d3f07138ee162824bb5def" Dec 09 16:39:38 crc kubenswrapper[4894]: E1209 16:39:38.647682 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0bb033a956a7bc057898f56b56faf95ed01c561ce6d3f07138ee162824bb5def\": container with ID starting with 0bb033a956a7bc057898f56b56faf95ed01c561ce6d3f07138ee162824bb5def not found: ID does not exist" containerID="0bb033a956a7bc057898f56b56faf95ed01c561ce6d3f07138ee162824bb5def" Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.647743 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0bb033a956a7bc057898f56b56faf95ed01c561ce6d3f07138ee162824bb5def"} err="failed to get container status \"0bb033a956a7bc057898f56b56faf95ed01c561ce6d3f07138ee162824bb5def\": rpc error: code = NotFound desc = could not find container \"0bb033a956a7bc057898f56b56faf95ed01c561ce6d3f07138ee162824bb5def\": container with ID starting with 0bb033a956a7bc057898f56b56faf95ed01c561ce6d3f07138ee162824bb5def not found: ID does not exist" Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.879073 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-jhjxg"] Dec 09 16:39:38 crc kubenswrapper[4894]: I1209 16:39:38.888033 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-jhjxg"] Dec 09 16:39:40 crc kubenswrapper[4894]: I1209 16:39:40.120048 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a3d33af-2799-4c66-b963-3d6c70036854" path="/var/lib/kubelet/pods/2a3d33af-2799-4c66-b963-3d6c70036854/volumes" Dec 09 16:39:42 crc kubenswrapper[4894]: I1209 16:39:42.257867 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:39:42 crc kubenswrapper[4894]: I1209 16:39:42.258373 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:40:12 crc kubenswrapper[4894]: I1209 16:40:12.258781 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:40:12 crc kubenswrapper[4894]: I1209 16:40:12.259258 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:40:32 crc kubenswrapper[4894]: E1209 16:40:32.547481 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-horizontest:current-podified" Dec 09 16:40:32 crc kubenswrapper[4894]: E1209 16:40:32.548207 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:horizontest-tests-horizontest,Image:quay.io/podified-antelope-centos9/openstack-horizontest:current-podified,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:ADMIN_PASSWORD,Value:12345678,ValueFrom:nil,},EnvVar{Name:ADMIN_USERNAME,Value:admin,ValueFrom:nil,},EnvVar{Name:AUTH_URL,Value:https://keystone-public-openstack.apps-crc.testing,ValueFrom:nil,},EnvVar{Name:DASHBOARD_URL,Value:https://horizon-openstack.apps-crc.testing/,ValueFrom:nil,},EnvVar{Name:EXTRA_FLAG,Value:not pagination and test_users.py,ValueFrom:nil,},EnvVar{Name:FLAVOR_NAME,Value:m1.tiny,ValueFrom:nil,},EnvVar{Name:HORIZONTEST_DEBUG_MODE,Value:false,ValueFrom:nil,},EnvVar{Name:HORIZON_KEYS_FOLDER,Value:/etc/test_operator,ValueFrom:nil,},EnvVar{Name:HORIZON_LOGS_DIR_NAME,Value:horizon,ValueFrom:nil,},EnvVar{Name:HORIZON_REPO_BRANCH,Value:master,ValueFrom:nil,},EnvVar{Name:IMAGE_FILE,Value:/var/lib/horizontest/cirros-0.6.2-x86_64-disk.img,ValueFrom:nil,},EnvVar{Name:IMAGE_FILE_NAME,Value:cirros-0.6.2-x86_64-disk,ValueFrom:nil,},EnvVar{Name:IMAGE_URL,Value:http://download.cirros-cloud.net/0.6.2/cirros-0.6.2-x86_64-disk.img,ValueFrom:nil,},EnvVar{Name:PASSWORD,Value:horizontest,ValueFrom:nil,},EnvVar{Name:PROJECT_NAME,Value:horizontest,ValueFrom:nil,},EnvVar{Name:PROJECT_NAME_XPATH,Value://*[@class=\"context-project\"]//ancestor::ul,ValueFrom:nil,},EnvVar{Name:REPO_URL,Value:https://review.opendev.org/openstack/horizon,ValueFrom:nil,},EnvVar{Name:USER_NAME,Value:horizontest,ValueFrom:nil,},EnvVar{Name:USE_EXTERNAL_FILES,Value:True,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{2 0} {} 2 DecimalSI},memory: {{4294967296 0} {} 4Gi BinarySI},},Requests:ResourceList{cpu: {{1 0} {} 1 DecimalSI},memory: {{2147483648 0} {} 2Gi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:test-operator-ephemeral-workdir,ReadOnly:false,MountPath:/var/lib/horizontest,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-ephemeral-temporary,ReadOnly:false,MountPath:/tmp,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-logs,ReadOnly:false,MountPath:/var/lib/horizontest/external_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/var/lib/horizontest/.config/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:test-operator-clouds-config,ReadOnly:true,MountPath:/etc/openstack/clouds.yaml,SubPath:clouds.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:openstack-config-secret,ReadOnly:false,MountPath:/etc/openstack/secure.yaml,SubPath:secure.yaml,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ca-certs,ReadOnly:true,MountPath:/etc/pki/tls/certs/ca-bundle.trust.crt,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:ceph,ReadOnly:true,MountPath:/etc/ceph,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-qc64h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[NET_ADMIN NET_RAW],Drop:[],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42455,RunAsNonRoot:*false,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:*true,RunAsGroup:*42455,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod horizontest-tests-horizontest_openstack(6573694d-aa1d-47bc-bfcd-6d92f7416f1a): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 16:40:32 crc kubenswrapper[4894]: E1209 16:40:32.549727 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"horizontest-tests-horizontest\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/horizontest-tests-horizontest" podUID="6573694d-aa1d-47bc-bfcd-6d92f7416f1a" Dec 09 16:40:33 crc kubenswrapper[4894]: E1209 16:40:33.047125 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"horizontest-tests-horizontest\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-horizontest:current-podified\\\"\"" pod="openstack/horizontest-tests-horizontest" podUID="6573694d-aa1d-47bc-bfcd-6d92f7416f1a" Dec 09 16:40:42 crc kubenswrapper[4894]: I1209 16:40:42.258075 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:40:42 crc kubenswrapper[4894]: I1209 16:40:42.259881 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:40:42 crc kubenswrapper[4894]: I1209 16:40:42.260072 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 16:40:42 crc kubenswrapper[4894]: I1209 16:40:42.261089 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"09a30e3fcdd74db8135435e7d3692a70be84fb1b91c0d6ae45b63f5970ba4ceb"} pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 16:40:42 crc kubenswrapper[4894]: I1209 16:40:42.261243 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" containerID="cri-o://09a30e3fcdd74db8135435e7d3692a70be84fb1b91c0d6ae45b63f5970ba4ceb" gracePeriod=600 Dec 09 16:40:43 crc kubenswrapper[4894]: I1209 16:40:43.146696 4894 generic.go:334] "Generic (PLEG): container finished" podID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerID="09a30e3fcdd74db8135435e7d3692a70be84fb1b91c0d6ae45b63f5970ba4ceb" exitCode=0 Dec 09 16:40:43 crc kubenswrapper[4894]: I1209 16:40:43.146794 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerDied","Data":"09a30e3fcdd74db8135435e7d3692a70be84fb1b91c0d6ae45b63f5970ba4ceb"} Dec 09 16:40:43 crc kubenswrapper[4894]: I1209 16:40:43.147515 4894 scope.go:117] "RemoveContainer" containerID="5d057056e49520887c4ef87b11053e465da71185268e43e9a6a9ae05a1456754" Dec 09 16:40:44 crc kubenswrapper[4894]: I1209 16:40:44.158322 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerStarted","Data":"30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01"} Dec 09 16:40:48 crc kubenswrapper[4894]: I1209 16:40:48.202683 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"6573694d-aa1d-47bc-bfcd-6d92f7416f1a","Type":"ContainerStarted","Data":"717335a09781c1fcb204117fbaf785fbe729776f3390d8162633e58283eae765"} Dec 09 16:41:57 crc kubenswrapper[4894]: I1209 16:41:57.991191 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/horizontest-tests-horizontest" podStartSLOduration=74.495579941 podStartE2EDuration="2m22.991175558s" podCreationTimestamp="2025-12-09 16:39:35 +0000 UTC" firstStartedPulling="2025-12-09 16:39:37.456448175 +0000 UTC m=+4071.775658844" lastFinishedPulling="2025-12-09 16:40:45.952043782 +0000 UTC m=+4140.271254461" observedRunningTime="2025-12-09 16:40:48.228903228 +0000 UTC m=+4142.548113907" watchObservedRunningTime="2025-12-09 16:41:57.991175558 +0000 UTC m=+4212.310386227" Dec 09 16:41:58 crc kubenswrapper[4894]: I1209 16:41:58.000430 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-b2wqg"] Dec 09 16:41:58 crc kubenswrapper[4894]: E1209 16:41:58.000913 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a3d33af-2799-4c66-b963-3d6c70036854" containerName="extract-utilities" Dec 09 16:41:58 crc kubenswrapper[4894]: I1209 16:41:58.000935 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a3d33af-2799-4c66-b963-3d6c70036854" containerName="extract-utilities" Dec 09 16:41:58 crc kubenswrapper[4894]: E1209 16:41:58.000952 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a3d33af-2799-4c66-b963-3d6c70036854" containerName="extract-content" Dec 09 16:41:58 crc kubenswrapper[4894]: I1209 16:41:58.000961 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a3d33af-2799-4c66-b963-3d6c70036854" containerName="extract-content" Dec 09 16:41:58 crc kubenswrapper[4894]: E1209 16:41:58.000998 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a3d33af-2799-4c66-b963-3d6c70036854" containerName="registry-server" Dec 09 16:41:58 crc kubenswrapper[4894]: I1209 16:41:58.001007 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a3d33af-2799-4c66-b963-3d6c70036854" containerName="registry-server" Dec 09 16:41:58 crc kubenswrapper[4894]: I1209 16:41:58.001226 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a3d33af-2799-4c66-b963-3d6c70036854" containerName="registry-server" Dec 09 16:41:58 crc kubenswrapper[4894]: I1209 16:41:58.002984 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b2wqg" Dec 09 16:41:58 crc kubenswrapper[4894]: I1209 16:41:58.026432 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b2wqg"] Dec 09 16:41:58 crc kubenswrapper[4894]: I1209 16:41:58.107368 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c815508b-18ba-430e-830d-b26eea38c207-utilities\") pod \"community-operators-b2wqg\" (UID: \"c815508b-18ba-430e-830d-b26eea38c207\") " pod="openshift-marketplace/community-operators-b2wqg" Dec 09 16:41:58 crc kubenswrapper[4894]: I1209 16:41:58.107462 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c815508b-18ba-430e-830d-b26eea38c207-catalog-content\") pod \"community-operators-b2wqg\" (UID: \"c815508b-18ba-430e-830d-b26eea38c207\") " pod="openshift-marketplace/community-operators-b2wqg" Dec 09 16:41:58 crc kubenswrapper[4894]: I1209 16:41:58.107484 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qmm48\" (UniqueName: \"kubernetes.io/projected/c815508b-18ba-430e-830d-b26eea38c207-kube-api-access-qmm48\") pod \"community-operators-b2wqg\" (UID: \"c815508b-18ba-430e-830d-b26eea38c207\") " pod="openshift-marketplace/community-operators-b2wqg" Dec 09 16:41:58 crc kubenswrapper[4894]: I1209 16:41:58.209884 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c815508b-18ba-430e-830d-b26eea38c207-catalog-content\") pod \"community-operators-b2wqg\" (UID: \"c815508b-18ba-430e-830d-b26eea38c207\") " pod="openshift-marketplace/community-operators-b2wqg" Dec 09 16:41:58 crc kubenswrapper[4894]: I1209 16:41:58.210184 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qmm48\" (UniqueName: \"kubernetes.io/projected/c815508b-18ba-430e-830d-b26eea38c207-kube-api-access-qmm48\") pod \"community-operators-b2wqg\" (UID: \"c815508b-18ba-430e-830d-b26eea38c207\") " pod="openshift-marketplace/community-operators-b2wqg" Dec 09 16:41:58 crc kubenswrapper[4894]: I1209 16:41:58.210461 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c815508b-18ba-430e-830d-b26eea38c207-utilities\") pod \"community-operators-b2wqg\" (UID: \"c815508b-18ba-430e-830d-b26eea38c207\") " pod="openshift-marketplace/community-operators-b2wqg" Dec 09 16:41:58 crc kubenswrapper[4894]: I1209 16:41:58.210924 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c815508b-18ba-430e-830d-b26eea38c207-catalog-content\") pod \"community-operators-b2wqg\" (UID: \"c815508b-18ba-430e-830d-b26eea38c207\") " pod="openshift-marketplace/community-operators-b2wqg" Dec 09 16:41:58 crc kubenswrapper[4894]: I1209 16:41:58.210988 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c815508b-18ba-430e-830d-b26eea38c207-utilities\") pod \"community-operators-b2wqg\" (UID: \"c815508b-18ba-430e-830d-b26eea38c207\") " pod="openshift-marketplace/community-operators-b2wqg" Dec 09 16:41:58 crc kubenswrapper[4894]: I1209 16:41:58.234079 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qmm48\" (UniqueName: \"kubernetes.io/projected/c815508b-18ba-430e-830d-b26eea38c207-kube-api-access-qmm48\") pod \"community-operators-b2wqg\" (UID: \"c815508b-18ba-430e-830d-b26eea38c207\") " pod="openshift-marketplace/community-operators-b2wqg" Dec 09 16:41:58 crc kubenswrapper[4894]: I1209 16:41:58.326388 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b2wqg" Dec 09 16:41:58 crc kubenswrapper[4894]: I1209 16:41:58.806289 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-b2wqg"] Dec 09 16:41:58 crc kubenswrapper[4894]: I1209 16:41:58.932512 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b2wqg" event={"ID":"c815508b-18ba-430e-830d-b26eea38c207","Type":"ContainerStarted","Data":"45132978e220b312ce7deb92f232073b8f56f5cd8f1dc5cdf8387095406b5696"} Dec 09 16:41:59 crc kubenswrapper[4894]: I1209 16:41:59.943144 4894 generic.go:334] "Generic (PLEG): container finished" podID="c815508b-18ba-430e-830d-b26eea38c207" containerID="71ee35439de0017f17f6567ef5b5cf8432afbd57560f6ab11b4dae558a789667" exitCode=0 Dec 09 16:41:59 crc kubenswrapper[4894]: I1209 16:41:59.943254 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b2wqg" event={"ID":"c815508b-18ba-430e-830d-b26eea38c207","Type":"ContainerDied","Data":"71ee35439de0017f17f6567ef5b5cf8432afbd57560f6ab11b4dae558a789667"} Dec 09 16:42:01 crc kubenswrapper[4894]: I1209 16:42:01.962868 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b2wqg" event={"ID":"c815508b-18ba-430e-830d-b26eea38c207","Type":"ContainerStarted","Data":"652735473b71ba05c7e609f16b0c2b17b0b726cef42a640c10e3673d9f200f03"} Dec 09 16:42:03 crc kubenswrapper[4894]: I1209 16:42:03.997849 4894 generic.go:334] "Generic (PLEG): container finished" podID="c815508b-18ba-430e-830d-b26eea38c207" containerID="652735473b71ba05c7e609f16b0c2b17b0b726cef42a640c10e3673d9f200f03" exitCode=0 Dec 09 16:42:03 crc kubenswrapper[4894]: I1209 16:42:03.997948 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b2wqg" event={"ID":"c815508b-18ba-430e-830d-b26eea38c207","Type":"ContainerDied","Data":"652735473b71ba05c7e609f16b0c2b17b0b726cef42a640c10e3673d9f200f03"} Dec 09 16:42:07 crc kubenswrapper[4894]: I1209 16:42:07.046104 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b2wqg" event={"ID":"c815508b-18ba-430e-830d-b26eea38c207","Type":"ContainerStarted","Data":"dc36bf31b3da95b378fa2a200d505ca5eb40a5cd132cf65909dbc3dabbdfb460"} Dec 09 16:42:07 crc kubenswrapper[4894]: I1209 16:42:07.076779 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-b2wqg" podStartSLOduration=4.02112417 podStartE2EDuration="10.076758189s" podCreationTimestamp="2025-12-09 16:41:57 +0000 UTC" firstStartedPulling="2025-12-09 16:41:59.945099685 +0000 UTC m=+4214.264310354" lastFinishedPulling="2025-12-09 16:42:06.000733694 +0000 UTC m=+4220.319944373" observedRunningTime="2025-12-09 16:42:07.070856529 +0000 UTC m=+4221.390067218" watchObservedRunningTime="2025-12-09 16:42:07.076758189 +0000 UTC m=+4221.395968858" Dec 09 16:42:08 crc kubenswrapper[4894]: I1209 16:42:08.327346 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-b2wqg" Dec 09 16:42:08 crc kubenswrapper[4894]: I1209 16:42:08.328529 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-b2wqg" Dec 09 16:42:09 crc kubenswrapper[4894]: I1209 16:42:09.378138 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/community-operators-b2wqg" podUID="c815508b-18ba-430e-830d-b26eea38c207" containerName="registry-server" probeResult="failure" output=< Dec 09 16:42:09 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 09 16:42:09 crc kubenswrapper[4894]: > Dec 09 16:42:18 crc kubenswrapper[4894]: I1209 16:42:18.384109 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-b2wqg" Dec 09 16:42:18 crc kubenswrapper[4894]: I1209 16:42:18.430768 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-b2wqg" Dec 09 16:42:18 crc kubenswrapper[4894]: I1209 16:42:18.628327 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b2wqg"] Dec 09 16:42:20 crc kubenswrapper[4894]: I1209 16:42:20.161454 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-b2wqg" podUID="c815508b-18ba-430e-830d-b26eea38c207" containerName="registry-server" containerID="cri-o://dc36bf31b3da95b378fa2a200d505ca5eb40a5cd132cf65909dbc3dabbdfb460" gracePeriod=2 Dec 09 16:42:20 crc kubenswrapper[4894]: I1209 16:42:20.762957 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b2wqg" Dec 09 16:42:20 crc kubenswrapper[4894]: I1209 16:42:20.886514 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qmm48\" (UniqueName: \"kubernetes.io/projected/c815508b-18ba-430e-830d-b26eea38c207-kube-api-access-qmm48\") pod \"c815508b-18ba-430e-830d-b26eea38c207\" (UID: \"c815508b-18ba-430e-830d-b26eea38c207\") " Dec 09 16:42:20 crc kubenswrapper[4894]: I1209 16:42:20.886689 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c815508b-18ba-430e-830d-b26eea38c207-catalog-content\") pod \"c815508b-18ba-430e-830d-b26eea38c207\" (UID: \"c815508b-18ba-430e-830d-b26eea38c207\") " Dec 09 16:42:20 crc kubenswrapper[4894]: I1209 16:42:20.886780 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c815508b-18ba-430e-830d-b26eea38c207-utilities\") pod \"c815508b-18ba-430e-830d-b26eea38c207\" (UID: \"c815508b-18ba-430e-830d-b26eea38c207\") " Dec 09 16:42:20 crc kubenswrapper[4894]: I1209 16:42:20.887824 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c815508b-18ba-430e-830d-b26eea38c207-utilities" (OuterVolumeSpecName: "utilities") pod "c815508b-18ba-430e-830d-b26eea38c207" (UID: "c815508b-18ba-430e-830d-b26eea38c207"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:42:20 crc kubenswrapper[4894]: I1209 16:42:20.893221 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c815508b-18ba-430e-830d-b26eea38c207-kube-api-access-qmm48" (OuterVolumeSpecName: "kube-api-access-qmm48") pod "c815508b-18ba-430e-830d-b26eea38c207" (UID: "c815508b-18ba-430e-830d-b26eea38c207"). InnerVolumeSpecName "kube-api-access-qmm48". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:42:20 crc kubenswrapper[4894]: I1209 16:42:20.937794 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c815508b-18ba-430e-830d-b26eea38c207-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c815508b-18ba-430e-830d-b26eea38c207" (UID: "c815508b-18ba-430e-830d-b26eea38c207"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:42:20 crc kubenswrapper[4894]: I1209 16:42:20.992976 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c815508b-18ba-430e-830d-b26eea38c207-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:42:20 crc kubenswrapper[4894]: I1209 16:42:20.993021 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c815508b-18ba-430e-830d-b26eea38c207-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:42:20 crc kubenswrapper[4894]: I1209 16:42:20.993044 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qmm48\" (UniqueName: \"kubernetes.io/projected/c815508b-18ba-430e-830d-b26eea38c207-kube-api-access-qmm48\") on node \"crc\" DevicePath \"\"" Dec 09 16:42:21 crc kubenswrapper[4894]: I1209 16:42:21.175902 4894 generic.go:334] "Generic (PLEG): container finished" podID="c815508b-18ba-430e-830d-b26eea38c207" containerID="dc36bf31b3da95b378fa2a200d505ca5eb40a5cd132cf65909dbc3dabbdfb460" exitCode=0 Dec 09 16:42:21 crc kubenswrapper[4894]: I1209 16:42:21.175971 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b2wqg" event={"ID":"c815508b-18ba-430e-830d-b26eea38c207","Type":"ContainerDied","Data":"dc36bf31b3da95b378fa2a200d505ca5eb40a5cd132cf65909dbc3dabbdfb460"} Dec 09 16:42:21 crc kubenswrapper[4894]: I1209 16:42:21.176036 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-b2wqg" event={"ID":"c815508b-18ba-430e-830d-b26eea38c207","Type":"ContainerDied","Data":"45132978e220b312ce7deb92f232073b8f56f5cd8f1dc5cdf8387095406b5696"} Dec 09 16:42:21 crc kubenswrapper[4894]: I1209 16:42:21.176059 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-b2wqg" Dec 09 16:42:21 crc kubenswrapper[4894]: I1209 16:42:21.176070 4894 scope.go:117] "RemoveContainer" containerID="dc36bf31b3da95b378fa2a200d505ca5eb40a5cd132cf65909dbc3dabbdfb460" Dec 09 16:42:21 crc kubenswrapper[4894]: I1209 16:42:21.229681 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-b2wqg"] Dec 09 16:42:21 crc kubenswrapper[4894]: I1209 16:42:21.240689 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-b2wqg"] Dec 09 16:42:21 crc kubenswrapper[4894]: I1209 16:42:21.240786 4894 scope.go:117] "RemoveContainer" containerID="652735473b71ba05c7e609f16b0c2b17b0b726cef42a640c10e3673d9f200f03" Dec 09 16:42:21 crc kubenswrapper[4894]: I1209 16:42:21.267667 4894 scope.go:117] "RemoveContainer" containerID="71ee35439de0017f17f6567ef5b5cf8432afbd57560f6ab11b4dae558a789667" Dec 09 16:42:21 crc kubenswrapper[4894]: I1209 16:42:21.318587 4894 scope.go:117] "RemoveContainer" containerID="dc36bf31b3da95b378fa2a200d505ca5eb40a5cd132cf65909dbc3dabbdfb460" Dec 09 16:42:21 crc kubenswrapper[4894]: E1209 16:42:21.319152 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dc36bf31b3da95b378fa2a200d505ca5eb40a5cd132cf65909dbc3dabbdfb460\": container with ID starting with dc36bf31b3da95b378fa2a200d505ca5eb40a5cd132cf65909dbc3dabbdfb460 not found: ID does not exist" containerID="dc36bf31b3da95b378fa2a200d505ca5eb40a5cd132cf65909dbc3dabbdfb460" Dec 09 16:42:21 crc kubenswrapper[4894]: I1209 16:42:21.319199 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dc36bf31b3da95b378fa2a200d505ca5eb40a5cd132cf65909dbc3dabbdfb460"} err="failed to get container status \"dc36bf31b3da95b378fa2a200d505ca5eb40a5cd132cf65909dbc3dabbdfb460\": rpc error: code = NotFound desc = could not find container \"dc36bf31b3da95b378fa2a200d505ca5eb40a5cd132cf65909dbc3dabbdfb460\": container with ID starting with dc36bf31b3da95b378fa2a200d505ca5eb40a5cd132cf65909dbc3dabbdfb460 not found: ID does not exist" Dec 09 16:42:21 crc kubenswrapper[4894]: I1209 16:42:21.319231 4894 scope.go:117] "RemoveContainer" containerID="652735473b71ba05c7e609f16b0c2b17b0b726cef42a640c10e3673d9f200f03" Dec 09 16:42:21 crc kubenswrapper[4894]: E1209 16:42:21.319613 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"652735473b71ba05c7e609f16b0c2b17b0b726cef42a640c10e3673d9f200f03\": container with ID starting with 652735473b71ba05c7e609f16b0c2b17b0b726cef42a640c10e3673d9f200f03 not found: ID does not exist" containerID="652735473b71ba05c7e609f16b0c2b17b0b726cef42a640c10e3673d9f200f03" Dec 09 16:42:21 crc kubenswrapper[4894]: I1209 16:42:21.319661 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"652735473b71ba05c7e609f16b0c2b17b0b726cef42a640c10e3673d9f200f03"} err="failed to get container status \"652735473b71ba05c7e609f16b0c2b17b0b726cef42a640c10e3673d9f200f03\": rpc error: code = NotFound desc = could not find container \"652735473b71ba05c7e609f16b0c2b17b0b726cef42a640c10e3673d9f200f03\": container with ID starting with 652735473b71ba05c7e609f16b0c2b17b0b726cef42a640c10e3673d9f200f03 not found: ID does not exist" Dec 09 16:42:21 crc kubenswrapper[4894]: I1209 16:42:21.319683 4894 scope.go:117] "RemoveContainer" containerID="71ee35439de0017f17f6567ef5b5cf8432afbd57560f6ab11b4dae558a789667" Dec 09 16:42:21 crc kubenswrapper[4894]: E1209 16:42:21.319944 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71ee35439de0017f17f6567ef5b5cf8432afbd57560f6ab11b4dae558a789667\": container with ID starting with 71ee35439de0017f17f6567ef5b5cf8432afbd57560f6ab11b4dae558a789667 not found: ID does not exist" containerID="71ee35439de0017f17f6567ef5b5cf8432afbd57560f6ab11b4dae558a789667" Dec 09 16:42:21 crc kubenswrapper[4894]: I1209 16:42:21.319978 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71ee35439de0017f17f6567ef5b5cf8432afbd57560f6ab11b4dae558a789667"} err="failed to get container status \"71ee35439de0017f17f6567ef5b5cf8432afbd57560f6ab11b4dae558a789667\": rpc error: code = NotFound desc = could not find container \"71ee35439de0017f17f6567ef5b5cf8432afbd57560f6ab11b4dae558a789667\": container with ID starting with 71ee35439de0017f17f6567ef5b5cf8432afbd57560f6ab11b4dae558a789667 not found: ID does not exist" Dec 09 16:42:21 crc kubenswrapper[4894]: E1209 16:42:21.416882 4894 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc815508b_18ba_430e_830d_b26eea38c207.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc815508b_18ba_430e_830d_b26eea38c207.slice/crio-45132978e220b312ce7deb92f232073b8f56f5cd8f1dc5cdf8387095406b5696\": RecentStats: unable to find data in memory cache]" Dec 09 16:42:22 crc kubenswrapper[4894]: I1209 16:42:22.118553 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c815508b-18ba-430e-830d-b26eea38c207" path="/var/lib/kubelet/pods/c815508b-18ba-430e-830d-b26eea38c207/volumes" Dec 09 16:42:45 crc kubenswrapper[4894]: I1209 16:42:45.651204 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-vb8rf"] Dec 09 16:42:45 crc kubenswrapper[4894]: E1209 16:42:45.652971 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c815508b-18ba-430e-830d-b26eea38c207" containerName="extract-utilities" Dec 09 16:42:45 crc kubenswrapper[4894]: I1209 16:42:45.653067 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c815508b-18ba-430e-830d-b26eea38c207" containerName="extract-utilities" Dec 09 16:42:45 crc kubenswrapper[4894]: E1209 16:42:45.653145 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c815508b-18ba-430e-830d-b26eea38c207" containerName="extract-content" Dec 09 16:42:45 crc kubenswrapper[4894]: I1209 16:42:45.653206 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c815508b-18ba-430e-830d-b26eea38c207" containerName="extract-content" Dec 09 16:42:45 crc kubenswrapper[4894]: E1209 16:42:45.653266 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c815508b-18ba-430e-830d-b26eea38c207" containerName="registry-server" Dec 09 16:42:45 crc kubenswrapper[4894]: I1209 16:42:45.653323 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="c815508b-18ba-430e-830d-b26eea38c207" containerName="registry-server" Dec 09 16:42:45 crc kubenswrapper[4894]: I1209 16:42:45.653581 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="c815508b-18ba-430e-830d-b26eea38c207" containerName="registry-server" Dec 09 16:42:45 crc kubenswrapper[4894]: I1209 16:42:45.655202 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vb8rf" Dec 09 16:42:45 crc kubenswrapper[4894]: I1209 16:42:45.677143 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vb8rf"] Dec 09 16:42:45 crc kubenswrapper[4894]: I1209 16:42:45.767312 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92c441d4-48dc-4cda-9ea0-9011f704e2dc-utilities\") pod \"redhat-marketplace-vb8rf\" (UID: \"92c441d4-48dc-4cda-9ea0-9011f704e2dc\") " pod="openshift-marketplace/redhat-marketplace-vb8rf" Dec 09 16:42:45 crc kubenswrapper[4894]: I1209 16:42:45.767722 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92c441d4-48dc-4cda-9ea0-9011f704e2dc-catalog-content\") pod \"redhat-marketplace-vb8rf\" (UID: \"92c441d4-48dc-4cda-9ea0-9011f704e2dc\") " pod="openshift-marketplace/redhat-marketplace-vb8rf" Dec 09 16:42:45 crc kubenswrapper[4894]: I1209 16:42:45.767928 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vm98f\" (UniqueName: \"kubernetes.io/projected/92c441d4-48dc-4cda-9ea0-9011f704e2dc-kube-api-access-vm98f\") pod \"redhat-marketplace-vb8rf\" (UID: \"92c441d4-48dc-4cda-9ea0-9011f704e2dc\") " pod="openshift-marketplace/redhat-marketplace-vb8rf" Dec 09 16:42:45 crc kubenswrapper[4894]: I1209 16:42:45.869862 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92c441d4-48dc-4cda-9ea0-9011f704e2dc-utilities\") pod \"redhat-marketplace-vb8rf\" (UID: \"92c441d4-48dc-4cda-9ea0-9011f704e2dc\") " pod="openshift-marketplace/redhat-marketplace-vb8rf" Dec 09 16:42:45 crc kubenswrapper[4894]: I1209 16:42:45.870188 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92c441d4-48dc-4cda-9ea0-9011f704e2dc-catalog-content\") pod \"redhat-marketplace-vb8rf\" (UID: \"92c441d4-48dc-4cda-9ea0-9011f704e2dc\") " pod="openshift-marketplace/redhat-marketplace-vb8rf" Dec 09 16:42:45 crc kubenswrapper[4894]: I1209 16:42:45.870339 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92c441d4-48dc-4cda-9ea0-9011f704e2dc-utilities\") pod \"redhat-marketplace-vb8rf\" (UID: \"92c441d4-48dc-4cda-9ea0-9011f704e2dc\") " pod="openshift-marketplace/redhat-marketplace-vb8rf" Dec 09 16:42:45 crc kubenswrapper[4894]: I1209 16:42:45.870455 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vm98f\" (UniqueName: \"kubernetes.io/projected/92c441d4-48dc-4cda-9ea0-9011f704e2dc-kube-api-access-vm98f\") pod \"redhat-marketplace-vb8rf\" (UID: \"92c441d4-48dc-4cda-9ea0-9011f704e2dc\") " pod="openshift-marketplace/redhat-marketplace-vb8rf" Dec 09 16:42:45 crc kubenswrapper[4894]: I1209 16:42:45.870567 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92c441d4-48dc-4cda-9ea0-9011f704e2dc-catalog-content\") pod \"redhat-marketplace-vb8rf\" (UID: \"92c441d4-48dc-4cda-9ea0-9011f704e2dc\") " pod="openshift-marketplace/redhat-marketplace-vb8rf" Dec 09 16:42:45 crc kubenswrapper[4894]: I1209 16:42:45.893415 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vm98f\" (UniqueName: \"kubernetes.io/projected/92c441d4-48dc-4cda-9ea0-9011f704e2dc-kube-api-access-vm98f\") pod \"redhat-marketplace-vb8rf\" (UID: \"92c441d4-48dc-4cda-9ea0-9011f704e2dc\") " pod="openshift-marketplace/redhat-marketplace-vb8rf" Dec 09 16:42:45 crc kubenswrapper[4894]: I1209 16:42:45.976877 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vb8rf" Dec 09 16:42:46 crc kubenswrapper[4894]: I1209 16:42:46.469020 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-vb8rf"] Dec 09 16:42:47 crc kubenswrapper[4894]: I1209 16:42:47.473027 4894 generic.go:334] "Generic (PLEG): container finished" podID="92c441d4-48dc-4cda-9ea0-9011f704e2dc" containerID="c3fc1795ff78a88f7b7651e54d3d1b146b7f70ee8de0b7b5e952665303a61621" exitCode=0 Dec 09 16:42:47 crc kubenswrapper[4894]: I1209 16:42:47.473157 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vb8rf" event={"ID":"92c441d4-48dc-4cda-9ea0-9011f704e2dc","Type":"ContainerDied","Data":"c3fc1795ff78a88f7b7651e54d3d1b146b7f70ee8de0b7b5e952665303a61621"} Dec 09 16:42:47 crc kubenswrapper[4894]: I1209 16:42:47.473510 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vb8rf" event={"ID":"92c441d4-48dc-4cda-9ea0-9011f704e2dc","Type":"ContainerStarted","Data":"532b97d98b3a5e38f1a7d9f2c8a07906c26e28113ce3a835203b716003773219"} Dec 09 16:42:49 crc kubenswrapper[4894]: I1209 16:42:49.490234 4894 generic.go:334] "Generic (PLEG): container finished" podID="92c441d4-48dc-4cda-9ea0-9011f704e2dc" containerID="ae7e2ac2301a23c1a893e584d9b17fd6c22cfaac5ca6e0b5a41351f3eb1ce12c" exitCode=0 Dec 09 16:42:49 crc kubenswrapper[4894]: I1209 16:42:49.490316 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vb8rf" event={"ID":"92c441d4-48dc-4cda-9ea0-9011f704e2dc","Type":"ContainerDied","Data":"ae7e2ac2301a23c1a893e584d9b17fd6c22cfaac5ca6e0b5a41351f3eb1ce12c"} Dec 09 16:42:51 crc kubenswrapper[4894]: I1209 16:42:51.513133 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vb8rf" event={"ID":"92c441d4-48dc-4cda-9ea0-9011f704e2dc","Type":"ContainerStarted","Data":"3dbe47cd4f6aaa4e15ad21990a958fe8cc9966f05afacf904b167fdd6e83c894"} Dec 09 16:42:51 crc kubenswrapper[4894]: I1209 16:42:51.534962 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-vb8rf" podStartSLOduration=3.386186622 podStartE2EDuration="6.534932238s" podCreationTimestamp="2025-12-09 16:42:45 +0000 UTC" firstStartedPulling="2025-12-09 16:42:47.478364796 +0000 UTC m=+4261.797575465" lastFinishedPulling="2025-12-09 16:42:50.627110422 +0000 UTC m=+4264.946321081" observedRunningTime="2025-12-09 16:42:51.532300667 +0000 UTC m=+4265.851511336" watchObservedRunningTime="2025-12-09 16:42:51.534932238 +0000 UTC m=+4265.854142947" Dec 09 16:42:55 crc kubenswrapper[4894]: I1209 16:42:55.977813 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-vb8rf" Dec 09 16:42:55 crc kubenswrapper[4894]: I1209 16:42:55.978332 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-vb8rf" Dec 09 16:42:56 crc kubenswrapper[4894]: I1209 16:42:56.050155 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-vb8rf" Dec 09 16:42:56 crc kubenswrapper[4894]: I1209 16:42:56.621520 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-vb8rf" Dec 09 16:42:57 crc kubenswrapper[4894]: I1209 16:42:57.242705 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vb8rf"] Dec 09 16:42:58 crc kubenswrapper[4894]: I1209 16:42:58.589875 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-vb8rf" podUID="92c441d4-48dc-4cda-9ea0-9011f704e2dc" containerName="registry-server" containerID="cri-o://3dbe47cd4f6aaa4e15ad21990a958fe8cc9966f05afacf904b167fdd6e83c894" gracePeriod=2 Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.087962 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vb8rf" Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.263990 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vm98f\" (UniqueName: \"kubernetes.io/projected/92c441d4-48dc-4cda-9ea0-9011f704e2dc-kube-api-access-vm98f\") pod \"92c441d4-48dc-4cda-9ea0-9011f704e2dc\" (UID: \"92c441d4-48dc-4cda-9ea0-9011f704e2dc\") " Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.264160 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92c441d4-48dc-4cda-9ea0-9011f704e2dc-utilities\") pod \"92c441d4-48dc-4cda-9ea0-9011f704e2dc\" (UID: \"92c441d4-48dc-4cda-9ea0-9011f704e2dc\") " Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.264429 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92c441d4-48dc-4cda-9ea0-9011f704e2dc-catalog-content\") pod \"92c441d4-48dc-4cda-9ea0-9011f704e2dc\" (UID: \"92c441d4-48dc-4cda-9ea0-9011f704e2dc\") " Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.269873 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92c441d4-48dc-4cda-9ea0-9011f704e2dc-kube-api-access-vm98f" (OuterVolumeSpecName: "kube-api-access-vm98f") pod "92c441d4-48dc-4cda-9ea0-9011f704e2dc" (UID: "92c441d4-48dc-4cda-9ea0-9011f704e2dc"). InnerVolumeSpecName "kube-api-access-vm98f". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.275002 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92c441d4-48dc-4cda-9ea0-9011f704e2dc-utilities" (OuterVolumeSpecName: "utilities") pod "92c441d4-48dc-4cda-9ea0-9011f704e2dc" (UID: "92c441d4-48dc-4cda-9ea0-9011f704e2dc"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.286250 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/92c441d4-48dc-4cda-9ea0-9011f704e2dc-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "92c441d4-48dc-4cda-9ea0-9011f704e2dc" (UID: "92c441d4-48dc-4cda-9ea0-9011f704e2dc"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.367361 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/92c441d4-48dc-4cda-9ea0-9011f704e2dc-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.367404 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/92c441d4-48dc-4cda-9ea0-9011f704e2dc-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.367422 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vm98f\" (UniqueName: \"kubernetes.io/projected/92c441d4-48dc-4cda-9ea0-9011f704e2dc-kube-api-access-vm98f\") on node \"crc\" DevicePath \"\"" Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.600906 4894 generic.go:334] "Generic (PLEG): container finished" podID="92c441d4-48dc-4cda-9ea0-9011f704e2dc" containerID="3dbe47cd4f6aaa4e15ad21990a958fe8cc9966f05afacf904b167fdd6e83c894" exitCode=0 Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.600959 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vb8rf" event={"ID":"92c441d4-48dc-4cda-9ea0-9011f704e2dc","Type":"ContainerDied","Data":"3dbe47cd4f6aaa4e15ad21990a958fe8cc9966f05afacf904b167fdd6e83c894"} Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.601012 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-vb8rf" event={"ID":"92c441d4-48dc-4cda-9ea0-9011f704e2dc","Type":"ContainerDied","Data":"532b97d98b3a5e38f1a7d9f2c8a07906c26e28113ce3a835203b716003773219"} Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.601010 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-vb8rf" Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.601031 4894 scope.go:117] "RemoveContainer" containerID="3dbe47cd4f6aaa4e15ad21990a958fe8cc9966f05afacf904b167fdd6e83c894" Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.640454 4894 scope.go:117] "RemoveContainer" containerID="ae7e2ac2301a23c1a893e584d9b17fd6c22cfaac5ca6e0b5a41351f3eb1ce12c" Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.652514 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-vb8rf"] Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.661136 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-vb8rf"] Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.668023 4894 scope.go:117] "RemoveContainer" containerID="c3fc1795ff78a88f7b7651e54d3d1b146b7f70ee8de0b7b5e952665303a61621" Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.709172 4894 scope.go:117] "RemoveContainer" containerID="3dbe47cd4f6aaa4e15ad21990a958fe8cc9966f05afacf904b167fdd6e83c894" Dec 09 16:42:59 crc kubenswrapper[4894]: E1209 16:42:59.710036 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3dbe47cd4f6aaa4e15ad21990a958fe8cc9966f05afacf904b167fdd6e83c894\": container with ID starting with 3dbe47cd4f6aaa4e15ad21990a958fe8cc9966f05afacf904b167fdd6e83c894 not found: ID does not exist" containerID="3dbe47cd4f6aaa4e15ad21990a958fe8cc9966f05afacf904b167fdd6e83c894" Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.710087 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3dbe47cd4f6aaa4e15ad21990a958fe8cc9966f05afacf904b167fdd6e83c894"} err="failed to get container status \"3dbe47cd4f6aaa4e15ad21990a958fe8cc9966f05afacf904b167fdd6e83c894\": rpc error: code = NotFound desc = could not find container \"3dbe47cd4f6aaa4e15ad21990a958fe8cc9966f05afacf904b167fdd6e83c894\": container with ID starting with 3dbe47cd4f6aaa4e15ad21990a958fe8cc9966f05afacf904b167fdd6e83c894 not found: ID does not exist" Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.710114 4894 scope.go:117] "RemoveContainer" containerID="ae7e2ac2301a23c1a893e584d9b17fd6c22cfaac5ca6e0b5a41351f3eb1ce12c" Dec 09 16:42:59 crc kubenswrapper[4894]: E1209 16:42:59.710790 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ae7e2ac2301a23c1a893e584d9b17fd6c22cfaac5ca6e0b5a41351f3eb1ce12c\": container with ID starting with ae7e2ac2301a23c1a893e584d9b17fd6c22cfaac5ca6e0b5a41351f3eb1ce12c not found: ID does not exist" containerID="ae7e2ac2301a23c1a893e584d9b17fd6c22cfaac5ca6e0b5a41351f3eb1ce12c" Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.710827 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ae7e2ac2301a23c1a893e584d9b17fd6c22cfaac5ca6e0b5a41351f3eb1ce12c"} err="failed to get container status \"ae7e2ac2301a23c1a893e584d9b17fd6c22cfaac5ca6e0b5a41351f3eb1ce12c\": rpc error: code = NotFound desc = could not find container \"ae7e2ac2301a23c1a893e584d9b17fd6c22cfaac5ca6e0b5a41351f3eb1ce12c\": container with ID starting with ae7e2ac2301a23c1a893e584d9b17fd6c22cfaac5ca6e0b5a41351f3eb1ce12c not found: ID does not exist" Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.710845 4894 scope.go:117] "RemoveContainer" containerID="c3fc1795ff78a88f7b7651e54d3d1b146b7f70ee8de0b7b5e952665303a61621" Dec 09 16:42:59 crc kubenswrapper[4894]: E1209 16:42:59.711329 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c3fc1795ff78a88f7b7651e54d3d1b146b7f70ee8de0b7b5e952665303a61621\": container with ID starting with c3fc1795ff78a88f7b7651e54d3d1b146b7f70ee8de0b7b5e952665303a61621 not found: ID does not exist" containerID="c3fc1795ff78a88f7b7651e54d3d1b146b7f70ee8de0b7b5e952665303a61621" Dec 09 16:42:59 crc kubenswrapper[4894]: I1209 16:42:59.711375 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c3fc1795ff78a88f7b7651e54d3d1b146b7f70ee8de0b7b5e952665303a61621"} err="failed to get container status \"c3fc1795ff78a88f7b7651e54d3d1b146b7f70ee8de0b7b5e952665303a61621\": rpc error: code = NotFound desc = could not find container \"c3fc1795ff78a88f7b7651e54d3d1b146b7f70ee8de0b7b5e952665303a61621\": container with ID starting with c3fc1795ff78a88f7b7651e54d3d1b146b7f70ee8de0b7b5e952665303a61621 not found: ID does not exist" Dec 09 16:43:00 crc kubenswrapper[4894]: I1209 16:43:00.125396 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92c441d4-48dc-4cda-9ea0-9011f704e2dc" path="/var/lib/kubelet/pods/92c441d4-48dc-4cda-9ea0-9011f704e2dc/volumes" Dec 09 16:43:06 crc kubenswrapper[4894]: I1209 16:43:06.680874 4894 generic.go:334] "Generic (PLEG): container finished" podID="6573694d-aa1d-47bc-bfcd-6d92f7416f1a" containerID="717335a09781c1fcb204117fbaf785fbe729776f3390d8162633e58283eae765" exitCode=0 Dec 09 16:43:06 crc kubenswrapper[4894]: I1209 16:43:06.681008 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"6573694d-aa1d-47bc-bfcd-6d92f7416f1a","Type":"ContainerDied","Data":"717335a09781c1fcb204117fbaf785fbe729776f3390d8162633e58283eae765"} Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.083623 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.269154 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qc64h\" (UniqueName: \"kubernetes.io/projected/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-kube-api-access-qc64h\") pod \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.269221 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-openstack-config-secret\") pod \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.269251 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-ca-certs\") pod \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.269314 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-logs\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.269351 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-ceph\") pod \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.269418 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-test-operator-ephemeral-temporary\") pod \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.269477 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-test-operator-ephemeral-workdir\") pod \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.269567 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-test-operator-clouds-config\") pod \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\" (UID: \"6573694d-aa1d-47bc-bfcd-6d92f7416f1a\") " Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.269913 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-test-operator-ephemeral-temporary" (OuterVolumeSpecName: "test-operator-ephemeral-temporary") pod "6573694d-aa1d-47bc-bfcd-6d92f7416f1a" (UID: "6573694d-aa1d-47bc-bfcd-6d92f7416f1a"). InnerVolumeSpecName "test-operator-ephemeral-temporary". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.270073 4894 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-temporary\" (UniqueName: \"kubernetes.io/empty-dir/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-test-operator-ephemeral-temporary\") on node \"crc\" DevicePath \"\"" Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.276754 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-kube-api-access-qc64h" (OuterVolumeSpecName: "kube-api-access-qc64h") pod "6573694d-aa1d-47bc-bfcd-6d92f7416f1a" (UID: "6573694d-aa1d-47bc-bfcd-6d92f7416f1a"). InnerVolumeSpecName "kube-api-access-qc64h". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.277187 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "test-operator-logs") pod "6573694d-aa1d-47bc-bfcd-6d92f7416f1a" (UID: "6573694d-aa1d-47bc-bfcd-6d92f7416f1a"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.279613 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-ceph" (OuterVolumeSpecName: "ceph") pod "6573694d-aa1d-47bc-bfcd-6d92f7416f1a" (UID: "6573694d-aa1d-47bc-bfcd-6d92f7416f1a"). InnerVolumeSpecName "ceph". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.327325 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "6573694d-aa1d-47bc-bfcd-6d92f7416f1a" (UID: "6573694d-aa1d-47bc-bfcd-6d92f7416f1a"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.345728 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-ca-certs" (OuterVolumeSpecName: "ca-certs") pod "6573694d-aa1d-47bc-bfcd-6d92f7416f1a" (UID: "6573694d-aa1d-47bc-bfcd-6d92f7416f1a"). InnerVolumeSpecName "ca-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.345970 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-test-operator-clouds-config" (OuterVolumeSpecName: "test-operator-clouds-config") pod "6573694d-aa1d-47bc-bfcd-6d92f7416f1a" (UID: "6573694d-aa1d-47bc-bfcd-6d92f7416f1a"). InnerVolumeSpecName "test-operator-clouds-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.372027 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qc64h\" (UniqueName: \"kubernetes.io/projected/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-kube-api-access-qc64h\") on node \"crc\" DevicePath \"\"" Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.372074 4894 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.372085 4894 reconciler_common.go:293] "Volume detached for volume \"ca-certs\" (UniqueName: \"kubernetes.io/secret/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-ca-certs\") on node \"crc\" DevicePath \"\"" Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.372123 4894 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.372138 4894 reconciler_common.go:293] "Volume detached for volume \"ceph\" (UniqueName: \"kubernetes.io/secret/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-ceph\") on node \"crc\" DevicePath \"\"" Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.372147 4894 reconciler_common.go:293] "Volume detached for volume \"test-operator-clouds-config\" (UniqueName: \"kubernetes.io/configmap/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-test-operator-clouds-config\") on node \"crc\" DevicePath \"\"" Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.396607 4894 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.476135 4894 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.492530 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-test-operator-ephemeral-workdir" (OuterVolumeSpecName: "test-operator-ephemeral-workdir") pod "6573694d-aa1d-47bc-bfcd-6d92f7416f1a" (UID: "6573694d-aa1d-47bc-bfcd-6d92f7416f1a"). InnerVolumeSpecName "test-operator-ephemeral-workdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.577850 4894 reconciler_common.go:293] "Volume detached for volume \"test-operator-ephemeral-workdir\" (UniqueName: \"kubernetes.io/empty-dir/6573694d-aa1d-47bc-bfcd-6d92f7416f1a-test-operator-ephemeral-workdir\") on node \"crc\" DevicePath \"\"" Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.701518 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/horizontest-tests-horizontest" event={"ID":"6573694d-aa1d-47bc-bfcd-6d92f7416f1a","Type":"ContainerDied","Data":"9e4a8e993e89b3d75990dbe9ceadf9c0516bd528b359cbe476bcadc3569c819b"} Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.701569 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9e4a8e993e89b3d75990dbe9ceadf9c0516bd528b359cbe476bcadc3569c819b" Dec 09 16:43:08 crc kubenswrapper[4894]: I1209 16:43:08.701603 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/horizontest-tests-horizontest" Dec 09 16:43:12 crc kubenswrapper[4894]: I1209 16:43:12.257401 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:43:12 crc kubenswrapper[4894]: I1209 16:43:12.259786 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:43:13 crc kubenswrapper[4894]: I1209 16:43:13.450383 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest"] Dec 09 16:43:13 crc kubenswrapper[4894]: E1209 16:43:13.451200 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92c441d4-48dc-4cda-9ea0-9011f704e2dc" containerName="extract-utilities" Dec 09 16:43:13 crc kubenswrapper[4894]: I1209 16:43:13.451217 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="92c441d4-48dc-4cda-9ea0-9011f704e2dc" containerName="extract-utilities" Dec 09 16:43:13 crc kubenswrapper[4894]: E1209 16:43:13.451244 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92c441d4-48dc-4cda-9ea0-9011f704e2dc" containerName="extract-content" Dec 09 16:43:13 crc kubenswrapper[4894]: I1209 16:43:13.451252 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="92c441d4-48dc-4cda-9ea0-9011f704e2dc" containerName="extract-content" Dec 09 16:43:13 crc kubenswrapper[4894]: E1209 16:43:13.451271 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6573694d-aa1d-47bc-bfcd-6d92f7416f1a" containerName="horizontest-tests-horizontest" Dec 09 16:43:13 crc kubenswrapper[4894]: I1209 16:43:13.451280 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="6573694d-aa1d-47bc-bfcd-6d92f7416f1a" containerName="horizontest-tests-horizontest" Dec 09 16:43:13 crc kubenswrapper[4894]: E1209 16:43:13.451307 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92c441d4-48dc-4cda-9ea0-9011f704e2dc" containerName="registry-server" Dec 09 16:43:13 crc kubenswrapper[4894]: I1209 16:43:13.451315 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="92c441d4-48dc-4cda-9ea0-9011f704e2dc" containerName="registry-server" Dec 09 16:43:13 crc kubenswrapper[4894]: I1209 16:43:13.451567 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="92c441d4-48dc-4cda-9ea0-9011f704e2dc" containerName="registry-server" Dec 09 16:43:13 crc kubenswrapper[4894]: I1209 16:43:13.451589 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="6573694d-aa1d-47bc-bfcd-6d92f7416f1a" containerName="horizontest-tests-horizontest" Dec 09 16:43:13 crc kubenswrapper[4894]: I1209 16:43:13.452407 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Dec 09 16:43:13 crc kubenswrapper[4894]: I1209 16:43:13.470213 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest"] Dec 09 16:43:13 crc kubenswrapper[4894]: I1209 16:43:13.579526 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccdln\" (UniqueName: \"kubernetes.io/projected/7f860d47-2922-4b88-962d-3d1d80564221-kube-api-access-ccdln\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"7f860d47-2922-4b88-962d-3d1d80564221\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Dec 09 16:43:13 crc kubenswrapper[4894]: I1209 16:43:13.579995 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"7f860d47-2922-4b88-962d-3d1d80564221\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Dec 09 16:43:13 crc kubenswrapper[4894]: I1209 16:43:13.683453 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"7f860d47-2922-4b88-962d-3d1d80564221\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Dec 09 16:43:13 crc kubenswrapper[4894]: I1209 16:43:13.683790 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccdln\" (UniqueName: \"kubernetes.io/projected/7f860d47-2922-4b88-962d-3d1d80564221-kube-api-access-ccdln\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"7f860d47-2922-4b88-962d-3d1d80564221\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Dec 09 16:43:13 crc kubenswrapper[4894]: I1209 16:43:13.684194 4894 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"7f860d47-2922-4b88-962d-3d1d80564221\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Dec 09 16:43:13 crc kubenswrapper[4894]: I1209 16:43:13.709899 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccdln\" (UniqueName: \"kubernetes.io/projected/7f860d47-2922-4b88-962d-3d1d80564221-kube-api-access-ccdln\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"7f860d47-2922-4b88-962d-3d1d80564221\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Dec 09 16:43:13 crc kubenswrapper[4894]: I1209 16:43:13.722515 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"test-operator-logs-pod-horizontest-horizontest-tests-horizontest\" (UID: \"7f860d47-2922-4b88-962d-3d1d80564221\") " pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Dec 09 16:43:13 crc kubenswrapper[4894]: I1209 16:43:13.783288 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" Dec 09 16:43:13 crc kubenswrapper[4894]: E1209 16:43:13.783449 4894 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 09 16:43:14 crc kubenswrapper[4894]: I1209 16:43:14.244889 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest"] Dec 09 16:43:14 crc kubenswrapper[4894]: W1209 16:43:14.717699 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7f860d47_2922_4b88_962d_3d1d80564221.slice/crio-3ff48495ddb4514137c707de88f8ec0b47fad387ffc9c703ea90ae120e48a095 WatchSource:0}: Error finding container 3ff48495ddb4514137c707de88f8ec0b47fad387ffc9c703ea90ae120e48a095: Status 404 returned error can't find the container with id 3ff48495ddb4514137c707de88f8ec0b47fad387ffc9c703ea90ae120e48a095 Dec 09 16:43:14 crc kubenswrapper[4894]: E1209 16:43:14.721611 4894 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 09 16:43:14 crc kubenswrapper[4894]: I1209 16:43:14.724428 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 16:43:14 crc kubenswrapper[4894]: I1209 16:43:14.761127 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" event={"ID":"7f860d47-2922-4b88-962d-3d1d80564221","Type":"ContainerStarted","Data":"3ff48495ddb4514137c707de88f8ec0b47fad387ffc9c703ea90ae120e48a095"} Dec 09 16:43:17 crc kubenswrapper[4894]: E1209 16:43:17.391987 4894 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 09 16:43:17 crc kubenswrapper[4894]: I1209 16:43:17.793521 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" event={"ID":"7f860d47-2922-4b88-962d-3d1d80564221","Type":"ContainerStarted","Data":"49d50b0b3b9a57d739ad280d27399f6d57e37962f64e56a1c1b3b3e3174c0d83"} Dec 09 16:43:17 crc kubenswrapper[4894]: E1209 16:43:17.795225 4894 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 09 16:43:17 crc kubenswrapper[4894]: I1209 16:43:17.809293 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/test-operator-logs-pod-horizontest-horizontest-tests-horizontest" podStartSLOduration=2.141053475 podStartE2EDuration="4.809271364s" podCreationTimestamp="2025-12-09 16:43:13 +0000 UTC" firstStartedPulling="2025-12-09 16:43:14.723537792 +0000 UTC m=+4289.042748501" lastFinishedPulling="2025-12-09 16:43:17.391755681 +0000 UTC m=+4291.710966390" observedRunningTime="2025-12-09 16:43:17.805671085 +0000 UTC m=+4292.124881804" watchObservedRunningTime="2025-12-09 16:43:17.809271364 +0000 UTC m=+4292.128482033" Dec 09 16:43:18 crc kubenswrapper[4894]: E1209 16:43:18.801728 4894 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 09 16:43:42 crc kubenswrapper[4894]: I1209 16:43:42.257276 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:43:42 crc kubenswrapper[4894]: I1209 16:43:42.257929 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:43:54 crc kubenswrapper[4894]: I1209 16:43:54.771195 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pn9dz/must-gather-xjnpp"] Dec 09 16:43:54 crc kubenswrapper[4894]: I1209 16:43:54.773703 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pn9dz/must-gather-xjnpp" Dec 09 16:43:54 crc kubenswrapper[4894]: I1209 16:43:54.775894 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-pn9dz"/"default-dockercfg-hv88s" Dec 09 16:43:54 crc kubenswrapper[4894]: I1209 16:43:54.776148 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-pn9dz"/"kube-root-ca.crt" Dec 09 16:43:54 crc kubenswrapper[4894]: I1209 16:43:54.776271 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-pn9dz"/"openshift-service-ca.crt" Dec 09 16:43:54 crc kubenswrapper[4894]: I1209 16:43:54.782269 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-pn9dz/must-gather-xjnpp"] Dec 09 16:43:54 crc kubenswrapper[4894]: I1209 16:43:54.868082 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ft5f4\" (UniqueName: \"kubernetes.io/projected/803d496d-7f84-4100-9f68-e2413ff22a54-kube-api-access-ft5f4\") pod \"must-gather-xjnpp\" (UID: \"803d496d-7f84-4100-9f68-e2413ff22a54\") " pod="openshift-must-gather-pn9dz/must-gather-xjnpp" Dec 09 16:43:54 crc kubenswrapper[4894]: I1209 16:43:54.868276 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/803d496d-7f84-4100-9f68-e2413ff22a54-must-gather-output\") pod \"must-gather-xjnpp\" (UID: \"803d496d-7f84-4100-9f68-e2413ff22a54\") " pod="openshift-must-gather-pn9dz/must-gather-xjnpp" Dec 09 16:43:54 crc kubenswrapper[4894]: I1209 16:43:54.971586 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/803d496d-7f84-4100-9f68-e2413ff22a54-must-gather-output\") pod \"must-gather-xjnpp\" (UID: \"803d496d-7f84-4100-9f68-e2413ff22a54\") " pod="openshift-must-gather-pn9dz/must-gather-xjnpp" Dec 09 16:43:54 crc kubenswrapper[4894]: I1209 16:43:54.971773 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ft5f4\" (UniqueName: \"kubernetes.io/projected/803d496d-7f84-4100-9f68-e2413ff22a54-kube-api-access-ft5f4\") pod \"must-gather-xjnpp\" (UID: \"803d496d-7f84-4100-9f68-e2413ff22a54\") " pod="openshift-must-gather-pn9dz/must-gather-xjnpp" Dec 09 16:43:54 crc kubenswrapper[4894]: I1209 16:43:54.972086 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/803d496d-7f84-4100-9f68-e2413ff22a54-must-gather-output\") pod \"must-gather-xjnpp\" (UID: \"803d496d-7f84-4100-9f68-e2413ff22a54\") " pod="openshift-must-gather-pn9dz/must-gather-xjnpp" Dec 09 16:43:54 crc kubenswrapper[4894]: I1209 16:43:54.996423 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ft5f4\" (UniqueName: \"kubernetes.io/projected/803d496d-7f84-4100-9f68-e2413ff22a54-kube-api-access-ft5f4\") pod \"must-gather-xjnpp\" (UID: \"803d496d-7f84-4100-9f68-e2413ff22a54\") " pod="openshift-must-gather-pn9dz/must-gather-xjnpp" Dec 09 16:43:55 crc kubenswrapper[4894]: I1209 16:43:55.093818 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pn9dz/must-gather-xjnpp" Dec 09 16:43:56 crc kubenswrapper[4894]: I1209 16:43:56.166397 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-pn9dz/must-gather-xjnpp"] Dec 09 16:43:56 crc kubenswrapper[4894]: I1209 16:43:56.176339 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pn9dz/must-gather-xjnpp" event={"ID":"803d496d-7f84-4100-9f68-e2413ff22a54","Type":"ContainerStarted","Data":"c5ef2bdc71a7a56d31c73930d34465263cabd33cfab58e80bdadae362e15c29e"} Dec 09 16:44:05 crc kubenswrapper[4894]: I1209 16:44:05.259597 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pn9dz/must-gather-xjnpp" event={"ID":"803d496d-7f84-4100-9f68-e2413ff22a54","Type":"ContainerStarted","Data":"4d585bb631c76873f173ca309f0f4ef760b09bc83ba0e5ad2fdda994558fd24b"} Dec 09 16:44:06 crc kubenswrapper[4894]: I1209 16:44:06.269380 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pn9dz/must-gather-xjnpp" event={"ID":"803d496d-7f84-4100-9f68-e2413ff22a54","Type":"ContainerStarted","Data":"ec73518edbce09c541073097d6977020ce9e53688c23c5782022b8bcf6012a9a"} Dec 09 16:44:06 crc kubenswrapper[4894]: I1209 16:44:06.287456 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pn9dz/must-gather-xjnpp" podStartSLOduration=5.12939535 podStartE2EDuration="12.287434232s" podCreationTimestamp="2025-12-09 16:43:54 +0000 UTC" firstStartedPulling="2025-12-09 16:43:56.16160191 +0000 UTC m=+4330.480812579" lastFinishedPulling="2025-12-09 16:44:03.319640792 +0000 UTC m=+4337.638851461" observedRunningTime="2025-12-09 16:44:06.283110674 +0000 UTC m=+4340.602321343" watchObservedRunningTime="2025-12-09 16:44:06.287434232 +0000 UTC m=+4340.606644901" Dec 09 16:44:06 crc kubenswrapper[4894]: E1209 16:44:06.862946 4894 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.64:37548->38.102.83.64:33645: write tcp 38.102.83.64:37548->38.102.83.64:33645: write: broken pipe Dec 09 16:44:08 crc kubenswrapper[4894]: I1209 16:44:08.953698 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pn9dz/crc-debug-b6x66"] Dec 09 16:44:08 crc kubenswrapper[4894]: I1209 16:44:08.955445 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pn9dz/crc-debug-b6x66" Dec 09 16:44:09 crc kubenswrapper[4894]: I1209 16:44:09.079521 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kk8b8\" (UniqueName: \"kubernetes.io/projected/7925ca9f-d2aa-4b5b-9018-2aae0e5bf728-kube-api-access-kk8b8\") pod \"crc-debug-b6x66\" (UID: \"7925ca9f-d2aa-4b5b-9018-2aae0e5bf728\") " pod="openshift-must-gather-pn9dz/crc-debug-b6x66" Dec 09 16:44:09 crc kubenswrapper[4894]: I1209 16:44:09.080341 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7925ca9f-d2aa-4b5b-9018-2aae0e5bf728-host\") pod \"crc-debug-b6x66\" (UID: \"7925ca9f-d2aa-4b5b-9018-2aae0e5bf728\") " pod="openshift-must-gather-pn9dz/crc-debug-b6x66" Dec 09 16:44:09 crc kubenswrapper[4894]: I1209 16:44:09.183077 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7925ca9f-d2aa-4b5b-9018-2aae0e5bf728-host\") pod \"crc-debug-b6x66\" (UID: \"7925ca9f-d2aa-4b5b-9018-2aae0e5bf728\") " pod="openshift-must-gather-pn9dz/crc-debug-b6x66" Dec 09 16:44:09 crc kubenswrapper[4894]: I1209 16:44:09.183245 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kk8b8\" (UniqueName: \"kubernetes.io/projected/7925ca9f-d2aa-4b5b-9018-2aae0e5bf728-kube-api-access-kk8b8\") pod \"crc-debug-b6x66\" (UID: \"7925ca9f-d2aa-4b5b-9018-2aae0e5bf728\") " pod="openshift-must-gather-pn9dz/crc-debug-b6x66" Dec 09 16:44:09 crc kubenswrapper[4894]: I1209 16:44:09.184204 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7925ca9f-d2aa-4b5b-9018-2aae0e5bf728-host\") pod \"crc-debug-b6x66\" (UID: \"7925ca9f-d2aa-4b5b-9018-2aae0e5bf728\") " pod="openshift-must-gather-pn9dz/crc-debug-b6x66" Dec 09 16:44:09 crc kubenswrapper[4894]: I1209 16:44:09.203476 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kk8b8\" (UniqueName: \"kubernetes.io/projected/7925ca9f-d2aa-4b5b-9018-2aae0e5bf728-kube-api-access-kk8b8\") pod \"crc-debug-b6x66\" (UID: \"7925ca9f-d2aa-4b5b-9018-2aae0e5bf728\") " pod="openshift-must-gather-pn9dz/crc-debug-b6x66" Dec 09 16:44:09 crc kubenswrapper[4894]: I1209 16:44:09.275041 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pn9dz/crc-debug-b6x66" Dec 09 16:44:09 crc kubenswrapper[4894]: W1209 16:44:09.308619 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7925ca9f_d2aa_4b5b_9018_2aae0e5bf728.slice/crio-7ff389a6dc3c360ab1ba89440a374cab36a37e52fa8349f6b73e0876cac8866e WatchSource:0}: Error finding container 7ff389a6dc3c360ab1ba89440a374cab36a37e52fa8349f6b73e0876cac8866e: Status 404 returned error can't find the container with id 7ff389a6dc3c360ab1ba89440a374cab36a37e52fa8349f6b73e0876cac8866e Dec 09 16:44:10 crc kubenswrapper[4894]: I1209 16:44:10.317154 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pn9dz/crc-debug-b6x66" event={"ID":"7925ca9f-d2aa-4b5b-9018-2aae0e5bf728","Type":"ContainerStarted","Data":"7ff389a6dc3c360ab1ba89440a374cab36a37e52fa8349f6b73e0876cac8866e"} Dec 09 16:44:12 crc kubenswrapper[4894]: I1209 16:44:12.257571 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:44:12 crc kubenswrapper[4894]: I1209 16:44:12.258194 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:44:12 crc kubenswrapper[4894]: I1209 16:44:12.258247 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 16:44:12 crc kubenswrapper[4894]: I1209 16:44:12.259092 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01"} pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 16:44:12 crc kubenswrapper[4894]: I1209 16:44:12.259155 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" containerID="cri-o://30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" gracePeriod=600 Dec 09 16:44:12 crc kubenswrapper[4894]: E1209 16:44:12.380705 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:44:13 crc kubenswrapper[4894]: I1209 16:44:13.352861 4894 generic.go:334] "Generic (PLEG): container finished" podID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" exitCode=0 Dec 09 16:44:13 crc kubenswrapper[4894]: I1209 16:44:13.352917 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerDied","Data":"30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01"} Dec 09 16:44:13 crc kubenswrapper[4894]: I1209 16:44:13.352948 4894 scope.go:117] "RemoveContainer" containerID="09a30e3fcdd74db8135435e7d3692a70be84fb1b91c0d6ae45b63f5970ba4ceb" Dec 09 16:44:13 crc kubenswrapper[4894]: I1209 16:44:13.353561 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:44:13 crc kubenswrapper[4894]: E1209 16:44:13.354074 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:44:24 crc kubenswrapper[4894]: I1209 16:44:24.106927 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:44:24 crc kubenswrapper[4894]: E1209 16:44:24.108013 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:44:32 crc kubenswrapper[4894]: E1209 16:44:32.106910 4894 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 09 16:44:37 crc kubenswrapper[4894]: I1209 16:44:37.107078 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:44:37 crc kubenswrapper[4894]: E1209 16:44:37.107829 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:44:37 crc kubenswrapper[4894]: E1209 16:44:37.379860 4894 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296" Dec 09 16:44:37 crc kubenswrapper[4894]: E1209 16:44:37.380180 4894 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:container-00,Image:quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296,Command:[chroot /host bash -c echo 'TOOLBOX_NAME=toolbox-osp' > /root/.toolboxrc ; rm -rf \"/var/tmp/sos-osp\" && mkdir -p \"/var/tmp/sos-osp\" && sudo podman rm --force toolbox-osp; sudo --preserve-env podman pull --authfile /var/lib/kubelet/config.json registry.redhat.io/rhel9/support-tools && toolbox sos report --batch --all-logs --only-plugins block,cifs,crio,devicemapper,devices,firewall_tables,firewalld,iscsi,lvm2,memory,multipath,nfs,nis,nvme,podman,process,processor,selinux,scsi,udev,logs,crypto --tmp-dir=\"/var/tmp/sos-osp\" && if [[ \"$(ls /var/log/pods/*/{*.log.*,*/*.log.*} 2>/dev/null)\" != '' ]]; then tar --ignore-failed-read --warning=no-file-changed -cJf \"/var/tmp/sos-osp/podlogs.tar.xz\" --transform 's,^,podlogs/,' /var/log/pods/*/{*.log.*,*/*.log.*} || true; fi],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:TMOUT,Value:900,ValueFrom:nil,},EnvVar{Name:HOST,Value:/host,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:host,ReadOnly:false,MountPath:/host,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-kk8b8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:*true,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod crc-debug-b6x66_openshift-must-gather-pn9dz(7925ca9f-d2aa-4b5b-9018-2aae0e5bf728): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Dec 09 16:44:37 crc kubenswrapper[4894]: E1209 16:44:37.381466 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openshift-must-gather-pn9dz/crc-debug-b6x66" podUID="7925ca9f-d2aa-4b5b-9018-2aae0e5bf728" Dec 09 16:44:37 crc kubenswrapper[4894]: E1209 16:44:37.609352 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"container-00\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:6ab858aed98e4fe57e6b144da8e90ad5d6698bb4cc5521206f5c05809f0f9296\\\"\"" pod="openshift-must-gather-pn9dz/crc-debug-b6x66" podUID="7925ca9f-d2aa-4b5b-9018-2aae0e5bf728" Dec 09 16:44:52 crc kubenswrapper[4894]: I1209 16:44:52.107432 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:44:52 crc kubenswrapper[4894]: E1209 16:44:52.108333 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:44:52 crc kubenswrapper[4894]: I1209 16:44:52.755251 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pn9dz/crc-debug-b6x66" event={"ID":"7925ca9f-d2aa-4b5b-9018-2aae0e5bf728","Type":"ContainerStarted","Data":"4b086c2991f9d8a7a0c710e444cb2ac0ad7d6dd29a053eff10716040119f9cc0"} Dec 09 16:44:52 crc kubenswrapper[4894]: I1209 16:44:52.776582 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-pn9dz/crc-debug-b6x66" podStartSLOduration=1.8794227110000001 podStartE2EDuration="44.77656192s" podCreationTimestamp="2025-12-09 16:44:08 +0000 UTC" firstStartedPulling="2025-12-09 16:44:09.311025911 +0000 UTC m=+4343.630236580" lastFinishedPulling="2025-12-09 16:44:52.20816512 +0000 UTC m=+4386.527375789" observedRunningTime="2025-12-09 16:44:52.766382922 +0000 UTC m=+4387.085593611" watchObservedRunningTime="2025-12-09 16:44:52.77656192 +0000 UTC m=+4387.095772589" Dec 09 16:45:00 crc kubenswrapper[4894]: I1209 16:45:00.181963 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421645-7bx72"] Dec 09 16:45:00 crc kubenswrapper[4894]: I1209 16:45:00.184133 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-7bx72" Dec 09 16:45:00 crc kubenswrapper[4894]: I1209 16:45:00.187370 4894 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Dec 09 16:45:00 crc kubenswrapper[4894]: I1209 16:45:00.187481 4894 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Dec 09 16:45:00 crc kubenswrapper[4894]: I1209 16:45:00.188883 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421645-7bx72"] Dec 09 16:45:00 crc kubenswrapper[4894]: I1209 16:45:00.242038 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59a6bf0b-6271-4fb8-9b14-e9fafa5be334-config-volume\") pod \"collect-profiles-29421645-7bx72\" (UID: \"59a6bf0b-6271-4fb8-9b14-e9fafa5be334\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-7bx72" Dec 09 16:45:00 crc kubenswrapper[4894]: I1209 16:45:00.242132 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59a6bf0b-6271-4fb8-9b14-e9fafa5be334-secret-volume\") pod \"collect-profiles-29421645-7bx72\" (UID: \"59a6bf0b-6271-4fb8-9b14-e9fafa5be334\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-7bx72" Dec 09 16:45:00 crc kubenswrapper[4894]: I1209 16:45:00.242507 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kcqzz\" (UniqueName: \"kubernetes.io/projected/59a6bf0b-6271-4fb8-9b14-e9fafa5be334-kube-api-access-kcqzz\") pod \"collect-profiles-29421645-7bx72\" (UID: \"59a6bf0b-6271-4fb8-9b14-e9fafa5be334\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-7bx72" Dec 09 16:45:00 crc kubenswrapper[4894]: I1209 16:45:00.344387 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kcqzz\" (UniqueName: \"kubernetes.io/projected/59a6bf0b-6271-4fb8-9b14-e9fafa5be334-kube-api-access-kcqzz\") pod \"collect-profiles-29421645-7bx72\" (UID: \"59a6bf0b-6271-4fb8-9b14-e9fafa5be334\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-7bx72" Dec 09 16:45:00 crc kubenswrapper[4894]: I1209 16:45:00.344501 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59a6bf0b-6271-4fb8-9b14-e9fafa5be334-config-volume\") pod \"collect-profiles-29421645-7bx72\" (UID: \"59a6bf0b-6271-4fb8-9b14-e9fafa5be334\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-7bx72" Dec 09 16:45:00 crc kubenswrapper[4894]: I1209 16:45:00.344584 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59a6bf0b-6271-4fb8-9b14-e9fafa5be334-secret-volume\") pod \"collect-profiles-29421645-7bx72\" (UID: \"59a6bf0b-6271-4fb8-9b14-e9fafa5be334\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-7bx72" Dec 09 16:45:00 crc kubenswrapper[4894]: I1209 16:45:00.345448 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59a6bf0b-6271-4fb8-9b14-e9fafa5be334-config-volume\") pod \"collect-profiles-29421645-7bx72\" (UID: \"59a6bf0b-6271-4fb8-9b14-e9fafa5be334\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-7bx72" Dec 09 16:45:00 crc kubenswrapper[4894]: I1209 16:45:00.350954 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59a6bf0b-6271-4fb8-9b14-e9fafa5be334-secret-volume\") pod \"collect-profiles-29421645-7bx72\" (UID: \"59a6bf0b-6271-4fb8-9b14-e9fafa5be334\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-7bx72" Dec 09 16:45:00 crc kubenswrapper[4894]: I1209 16:45:00.361173 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kcqzz\" (UniqueName: \"kubernetes.io/projected/59a6bf0b-6271-4fb8-9b14-e9fafa5be334-kube-api-access-kcqzz\") pod \"collect-profiles-29421645-7bx72\" (UID: \"59a6bf0b-6271-4fb8-9b14-e9fafa5be334\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-7bx72" Dec 09 16:45:00 crc kubenswrapper[4894]: I1209 16:45:00.503434 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-7bx72" Dec 09 16:45:05 crc kubenswrapper[4894]: I1209 16:45:05.107360 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:45:05 crc kubenswrapper[4894]: E1209 16:45:05.108549 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:45:10 crc kubenswrapper[4894]: I1209 16:45:10.055658 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421645-7bx72"] Dec 09 16:45:10 crc kubenswrapper[4894]: W1209 16:45:10.057563 4894 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod59a6bf0b_6271_4fb8_9b14_e9fafa5be334.slice/crio-975d5dc065e184c4d5aed3019b391d1fed11103b5cee676dd4af9ccb1d7ee9ce WatchSource:0}: Error finding container 975d5dc065e184c4d5aed3019b391d1fed11103b5cee676dd4af9ccb1d7ee9ce: Status 404 returned error can't find the container with id 975d5dc065e184c4d5aed3019b391d1fed11103b5cee676dd4af9ccb1d7ee9ce Dec 09 16:45:10 crc kubenswrapper[4894]: I1209 16:45:10.926192 4894 generic.go:334] "Generic (PLEG): container finished" podID="59a6bf0b-6271-4fb8-9b14-e9fafa5be334" containerID="b8b50d96ed3d7a6124c6f2f4caf2e8233f6e85a87c41e76f2868e0d658840c55" exitCode=0 Dec 09 16:45:10 crc kubenswrapper[4894]: I1209 16:45:10.926289 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-7bx72" event={"ID":"59a6bf0b-6271-4fb8-9b14-e9fafa5be334","Type":"ContainerDied","Data":"b8b50d96ed3d7a6124c6f2f4caf2e8233f6e85a87c41e76f2868e0d658840c55"} Dec 09 16:45:10 crc kubenswrapper[4894]: I1209 16:45:10.926716 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-7bx72" event={"ID":"59a6bf0b-6271-4fb8-9b14-e9fafa5be334","Type":"ContainerStarted","Data":"975d5dc065e184c4d5aed3019b391d1fed11103b5cee676dd4af9ccb1d7ee9ce"} Dec 09 16:45:12 crc kubenswrapper[4894]: I1209 16:45:12.308456 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-7bx72" Dec 09 16:45:12 crc kubenswrapper[4894]: I1209 16:45:12.408310 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59a6bf0b-6271-4fb8-9b14-e9fafa5be334-config-volume\") pod \"59a6bf0b-6271-4fb8-9b14-e9fafa5be334\" (UID: \"59a6bf0b-6271-4fb8-9b14-e9fafa5be334\") " Dec 09 16:45:12 crc kubenswrapper[4894]: I1209 16:45:12.408480 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59a6bf0b-6271-4fb8-9b14-e9fafa5be334-secret-volume\") pod \"59a6bf0b-6271-4fb8-9b14-e9fafa5be334\" (UID: \"59a6bf0b-6271-4fb8-9b14-e9fafa5be334\") " Dec 09 16:45:12 crc kubenswrapper[4894]: I1209 16:45:12.408580 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kcqzz\" (UniqueName: \"kubernetes.io/projected/59a6bf0b-6271-4fb8-9b14-e9fafa5be334-kube-api-access-kcqzz\") pod \"59a6bf0b-6271-4fb8-9b14-e9fafa5be334\" (UID: \"59a6bf0b-6271-4fb8-9b14-e9fafa5be334\") " Dec 09 16:45:12 crc kubenswrapper[4894]: I1209 16:45:12.409113 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/59a6bf0b-6271-4fb8-9b14-e9fafa5be334-config-volume" (OuterVolumeSpecName: "config-volume") pod "59a6bf0b-6271-4fb8-9b14-e9fafa5be334" (UID: "59a6bf0b-6271-4fb8-9b14-e9fafa5be334"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Dec 09 16:45:12 crc kubenswrapper[4894]: I1209 16:45:12.414796 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/59a6bf0b-6271-4fb8-9b14-e9fafa5be334-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "59a6bf0b-6271-4fb8-9b14-e9fafa5be334" (UID: "59a6bf0b-6271-4fb8-9b14-e9fafa5be334"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Dec 09 16:45:12 crc kubenswrapper[4894]: I1209 16:45:12.415487 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/59a6bf0b-6271-4fb8-9b14-e9fafa5be334-kube-api-access-kcqzz" (OuterVolumeSpecName: "kube-api-access-kcqzz") pod "59a6bf0b-6271-4fb8-9b14-e9fafa5be334" (UID: "59a6bf0b-6271-4fb8-9b14-e9fafa5be334"). InnerVolumeSpecName "kube-api-access-kcqzz". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:45:12 crc kubenswrapper[4894]: I1209 16:45:12.511100 4894 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/59a6bf0b-6271-4fb8-9b14-e9fafa5be334-secret-volume\") on node \"crc\" DevicePath \"\"" Dec 09 16:45:12 crc kubenswrapper[4894]: I1209 16:45:12.511155 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kcqzz\" (UniqueName: \"kubernetes.io/projected/59a6bf0b-6271-4fb8-9b14-e9fafa5be334-kube-api-access-kcqzz\") on node \"crc\" DevicePath \"\"" Dec 09 16:45:12 crc kubenswrapper[4894]: I1209 16:45:12.511170 4894 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/59a6bf0b-6271-4fb8-9b14-e9fafa5be334-config-volume\") on node \"crc\" DevicePath \"\"" Dec 09 16:45:12 crc kubenswrapper[4894]: I1209 16:45:12.954610 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-7bx72" event={"ID":"59a6bf0b-6271-4fb8-9b14-e9fafa5be334","Type":"ContainerDied","Data":"975d5dc065e184c4d5aed3019b391d1fed11103b5cee676dd4af9ccb1d7ee9ce"} Dec 09 16:45:12 crc kubenswrapper[4894]: I1209 16:45:12.954992 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="975d5dc065e184c4d5aed3019b391d1fed11103b5cee676dd4af9ccb1d7ee9ce" Dec 09 16:45:12 crc kubenswrapper[4894]: I1209 16:45:12.955089 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29421645-7bx72" Dec 09 16:45:13 crc kubenswrapper[4894]: I1209 16:45:13.379753 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz"] Dec 09 16:45:13 crc kubenswrapper[4894]: I1209 16:45:13.390609 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29421600-t6btz"] Dec 09 16:45:14 crc kubenswrapper[4894]: I1209 16:45:14.121133 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="acd00528-8721-45e8-b085-c8dab62fbc7f" path="/var/lib/kubelet/pods/acd00528-8721-45e8-b085-c8dab62fbc7f/volumes" Dec 09 16:45:16 crc kubenswrapper[4894]: I1209 16:45:16.119969 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:45:16 crc kubenswrapper[4894]: E1209 16:45:16.120581 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:45:21 crc kubenswrapper[4894]: I1209 16:45:21.048186 4894 generic.go:334] "Generic (PLEG): container finished" podID="7925ca9f-d2aa-4b5b-9018-2aae0e5bf728" containerID="4b086c2991f9d8a7a0c710e444cb2ac0ad7d6dd29a053eff10716040119f9cc0" exitCode=0 Dec 09 16:45:21 crc kubenswrapper[4894]: I1209 16:45:21.048278 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pn9dz/crc-debug-b6x66" event={"ID":"7925ca9f-d2aa-4b5b-9018-2aae0e5bf728","Type":"ContainerDied","Data":"4b086c2991f9d8a7a0c710e444cb2ac0ad7d6dd29a053eff10716040119f9cc0"} Dec 09 16:45:22 crc kubenswrapper[4894]: I1209 16:45:22.185766 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pn9dz/crc-debug-b6x66" Dec 09 16:45:22 crc kubenswrapper[4894]: I1209 16:45:22.229814 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pn9dz/crc-debug-b6x66"] Dec 09 16:45:22 crc kubenswrapper[4894]: I1209 16:45:22.238833 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pn9dz/crc-debug-b6x66"] Dec 09 16:45:22 crc kubenswrapper[4894]: I1209 16:45:22.318108 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7925ca9f-d2aa-4b5b-9018-2aae0e5bf728-host\") pod \"7925ca9f-d2aa-4b5b-9018-2aae0e5bf728\" (UID: \"7925ca9f-d2aa-4b5b-9018-2aae0e5bf728\") " Dec 09 16:45:22 crc kubenswrapper[4894]: I1209 16:45:22.318181 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kk8b8\" (UniqueName: \"kubernetes.io/projected/7925ca9f-d2aa-4b5b-9018-2aae0e5bf728-kube-api-access-kk8b8\") pod \"7925ca9f-d2aa-4b5b-9018-2aae0e5bf728\" (UID: \"7925ca9f-d2aa-4b5b-9018-2aae0e5bf728\") " Dec 09 16:45:22 crc kubenswrapper[4894]: I1209 16:45:22.318224 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7925ca9f-d2aa-4b5b-9018-2aae0e5bf728-host" (OuterVolumeSpecName: "host") pod "7925ca9f-d2aa-4b5b-9018-2aae0e5bf728" (UID: "7925ca9f-d2aa-4b5b-9018-2aae0e5bf728"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 16:45:22 crc kubenswrapper[4894]: I1209 16:45:22.318851 4894 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/7925ca9f-d2aa-4b5b-9018-2aae0e5bf728-host\") on node \"crc\" DevicePath \"\"" Dec 09 16:45:22 crc kubenswrapper[4894]: I1209 16:45:22.324750 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7925ca9f-d2aa-4b5b-9018-2aae0e5bf728-kube-api-access-kk8b8" (OuterVolumeSpecName: "kube-api-access-kk8b8") pod "7925ca9f-d2aa-4b5b-9018-2aae0e5bf728" (UID: "7925ca9f-d2aa-4b5b-9018-2aae0e5bf728"). InnerVolumeSpecName "kube-api-access-kk8b8". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:45:22 crc kubenswrapper[4894]: I1209 16:45:22.421340 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kk8b8\" (UniqueName: \"kubernetes.io/projected/7925ca9f-d2aa-4b5b-9018-2aae0e5bf728-kube-api-access-kk8b8\") on node \"crc\" DevicePath \"\"" Dec 09 16:45:23 crc kubenswrapper[4894]: I1209 16:45:23.077843 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pn9dz/crc-debug-b6x66" Dec 09 16:45:23 crc kubenswrapper[4894]: I1209 16:45:23.077753 4894 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ff389a6dc3c360ab1ba89440a374cab36a37e52fa8349f6b73e0876cac8866e" Dec 09 16:45:23 crc kubenswrapper[4894]: I1209 16:45:23.415412 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-pn9dz/crc-debug-tfj88"] Dec 09 16:45:23 crc kubenswrapper[4894]: E1209 16:45:23.416104 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7925ca9f-d2aa-4b5b-9018-2aae0e5bf728" containerName="container-00" Dec 09 16:45:23 crc kubenswrapper[4894]: I1209 16:45:23.416118 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="7925ca9f-d2aa-4b5b-9018-2aae0e5bf728" containerName="container-00" Dec 09 16:45:23 crc kubenswrapper[4894]: E1209 16:45:23.416152 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="59a6bf0b-6271-4fb8-9b14-e9fafa5be334" containerName="collect-profiles" Dec 09 16:45:23 crc kubenswrapper[4894]: I1209 16:45:23.416158 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="59a6bf0b-6271-4fb8-9b14-e9fafa5be334" containerName="collect-profiles" Dec 09 16:45:23 crc kubenswrapper[4894]: I1209 16:45:23.416356 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="7925ca9f-d2aa-4b5b-9018-2aae0e5bf728" containerName="container-00" Dec 09 16:45:23 crc kubenswrapper[4894]: I1209 16:45:23.416372 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="59a6bf0b-6271-4fb8-9b14-e9fafa5be334" containerName="collect-profiles" Dec 09 16:45:23 crc kubenswrapper[4894]: I1209 16:45:23.417016 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pn9dz/crc-debug-tfj88" Dec 09 16:45:23 crc kubenswrapper[4894]: I1209 16:45:23.546149 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k29rc\" (UniqueName: \"kubernetes.io/projected/64eb0ebe-7743-4d2d-ba38-84ebb1b004e4-kube-api-access-k29rc\") pod \"crc-debug-tfj88\" (UID: \"64eb0ebe-7743-4d2d-ba38-84ebb1b004e4\") " pod="openshift-must-gather-pn9dz/crc-debug-tfj88" Dec 09 16:45:23 crc kubenswrapper[4894]: I1209 16:45:23.546298 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/64eb0ebe-7743-4d2d-ba38-84ebb1b004e4-host\") pod \"crc-debug-tfj88\" (UID: \"64eb0ebe-7743-4d2d-ba38-84ebb1b004e4\") " pod="openshift-must-gather-pn9dz/crc-debug-tfj88" Dec 09 16:45:23 crc kubenswrapper[4894]: I1209 16:45:23.648932 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/64eb0ebe-7743-4d2d-ba38-84ebb1b004e4-host\") pod \"crc-debug-tfj88\" (UID: \"64eb0ebe-7743-4d2d-ba38-84ebb1b004e4\") " pod="openshift-must-gather-pn9dz/crc-debug-tfj88" Dec 09 16:45:23 crc kubenswrapper[4894]: I1209 16:45:23.649122 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/64eb0ebe-7743-4d2d-ba38-84ebb1b004e4-host\") pod \"crc-debug-tfj88\" (UID: \"64eb0ebe-7743-4d2d-ba38-84ebb1b004e4\") " pod="openshift-must-gather-pn9dz/crc-debug-tfj88" Dec 09 16:45:23 crc kubenswrapper[4894]: I1209 16:45:23.649716 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k29rc\" (UniqueName: \"kubernetes.io/projected/64eb0ebe-7743-4d2d-ba38-84ebb1b004e4-kube-api-access-k29rc\") pod \"crc-debug-tfj88\" (UID: \"64eb0ebe-7743-4d2d-ba38-84ebb1b004e4\") " pod="openshift-must-gather-pn9dz/crc-debug-tfj88" Dec 09 16:45:23 crc kubenswrapper[4894]: I1209 16:45:23.671034 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k29rc\" (UniqueName: \"kubernetes.io/projected/64eb0ebe-7743-4d2d-ba38-84ebb1b004e4-kube-api-access-k29rc\") pod \"crc-debug-tfj88\" (UID: \"64eb0ebe-7743-4d2d-ba38-84ebb1b004e4\") " pod="openshift-must-gather-pn9dz/crc-debug-tfj88" Dec 09 16:45:23 crc kubenswrapper[4894]: I1209 16:45:23.734913 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pn9dz/crc-debug-tfj88" Dec 09 16:45:24 crc kubenswrapper[4894]: I1209 16:45:24.121228 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7925ca9f-d2aa-4b5b-9018-2aae0e5bf728" path="/var/lib/kubelet/pods/7925ca9f-d2aa-4b5b-9018-2aae0e5bf728/volumes" Dec 09 16:45:25 crc kubenswrapper[4894]: I1209 16:45:25.105431 4894 generic.go:334] "Generic (PLEG): container finished" podID="64eb0ebe-7743-4d2d-ba38-84ebb1b004e4" containerID="e127be05878d97153bcddd026390f1ea2dc6a10bc8073038ce7dafa0953af6a6" exitCode=1 Dec 09 16:45:25 crc kubenswrapper[4894]: I1209 16:45:25.105496 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pn9dz/crc-debug-tfj88" event={"ID":"64eb0ebe-7743-4d2d-ba38-84ebb1b004e4","Type":"ContainerDied","Data":"e127be05878d97153bcddd026390f1ea2dc6a10bc8073038ce7dafa0953af6a6"} Dec 09 16:45:25 crc kubenswrapper[4894]: I1209 16:45:25.105966 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pn9dz/crc-debug-tfj88" event={"ID":"64eb0ebe-7743-4d2d-ba38-84ebb1b004e4","Type":"ContainerStarted","Data":"7189f051665d578d48e56e0ec979b8ca631100e008173492510d246d04039315"} Dec 09 16:45:25 crc kubenswrapper[4894]: I1209 16:45:25.144491 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pn9dz/crc-debug-tfj88"] Dec 09 16:45:25 crc kubenswrapper[4894]: I1209 16:45:25.155570 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pn9dz/crc-debug-tfj88"] Dec 09 16:45:26 crc kubenswrapper[4894]: I1209 16:45:26.235976 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pn9dz/crc-debug-tfj88" Dec 09 16:45:26 crc kubenswrapper[4894]: I1209 16:45:26.304069 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/64eb0ebe-7743-4d2d-ba38-84ebb1b004e4-host\") pod \"64eb0ebe-7743-4d2d-ba38-84ebb1b004e4\" (UID: \"64eb0ebe-7743-4d2d-ba38-84ebb1b004e4\") " Dec 09 16:45:26 crc kubenswrapper[4894]: I1209 16:45:26.304194 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-k29rc\" (UniqueName: \"kubernetes.io/projected/64eb0ebe-7743-4d2d-ba38-84ebb1b004e4-kube-api-access-k29rc\") pod \"64eb0ebe-7743-4d2d-ba38-84ebb1b004e4\" (UID: \"64eb0ebe-7743-4d2d-ba38-84ebb1b004e4\") " Dec 09 16:45:26 crc kubenswrapper[4894]: I1209 16:45:26.304242 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/64eb0ebe-7743-4d2d-ba38-84ebb1b004e4-host" (OuterVolumeSpecName: "host") pod "64eb0ebe-7743-4d2d-ba38-84ebb1b004e4" (UID: "64eb0ebe-7743-4d2d-ba38-84ebb1b004e4"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Dec 09 16:45:26 crc kubenswrapper[4894]: I1209 16:45:26.304827 4894 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/64eb0ebe-7743-4d2d-ba38-84ebb1b004e4-host\") on node \"crc\" DevicePath \"\"" Dec 09 16:45:26 crc kubenswrapper[4894]: I1209 16:45:26.311071 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64eb0ebe-7743-4d2d-ba38-84ebb1b004e4-kube-api-access-k29rc" (OuterVolumeSpecName: "kube-api-access-k29rc") pod "64eb0ebe-7743-4d2d-ba38-84ebb1b004e4" (UID: "64eb0ebe-7743-4d2d-ba38-84ebb1b004e4"). InnerVolumeSpecName "kube-api-access-k29rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:45:26 crc kubenswrapper[4894]: I1209 16:45:26.407390 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-k29rc\" (UniqueName: \"kubernetes.io/projected/64eb0ebe-7743-4d2d-ba38-84ebb1b004e4-kube-api-access-k29rc\") on node \"crc\" DevicePath \"\"" Dec 09 16:45:27 crc kubenswrapper[4894]: I1209 16:45:27.124212 4894 scope.go:117] "RemoveContainer" containerID="e127be05878d97153bcddd026390f1ea2dc6a10bc8073038ce7dafa0953af6a6" Dec 09 16:45:27 crc kubenswrapper[4894]: I1209 16:45:27.124262 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pn9dz/crc-debug-tfj88" Dec 09 16:45:28 crc kubenswrapper[4894]: I1209 16:45:28.107211 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:45:28 crc kubenswrapper[4894]: E1209 16:45:28.108086 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:45:28 crc kubenswrapper[4894]: I1209 16:45:28.120034 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64eb0ebe-7743-4d2d-ba38-84ebb1b004e4" path="/var/lib/kubelet/pods/64eb0ebe-7743-4d2d-ba38-84ebb1b004e4/volumes" Dec 09 16:45:42 crc kubenswrapper[4894]: I1209 16:45:42.107786 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:45:42 crc kubenswrapper[4894]: E1209 16:45:42.108998 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:45:43 crc kubenswrapper[4894]: I1209 16:45:43.849664 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ansibletest-ansibletest_b56c2145-351c-42cf-8a4e-049c7825d527/ansibletest-ansibletest/0.log" Dec 09 16:45:44 crc kubenswrapper[4894]: I1209 16:45:44.790223 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5df7cc9c4d-kfjtd_fcc16386-61b0-4e99-a220-70f90f549e90/barbican-api/0.log" Dec 09 16:45:44 crc kubenswrapper[4894]: I1209 16:45:44.835879 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-5df7cc9c4d-kfjtd_fcc16386-61b0-4e99-a220-70f90f549e90/barbican-api-log/0.log" Dec 09 16:45:44 crc kubenswrapper[4894]: I1209 16:45:44.876751 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-67f4558b8d-v6cpx_b470d89f-2e08-4222-adc9-5dab929008ea/barbican-keystone-listener/0.log" Dec 09 16:45:44 crc kubenswrapper[4894]: I1209 16:45:44.968092 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-67f4558b8d-v6cpx_b470d89f-2e08-4222-adc9-5dab929008ea/barbican-keystone-listener-log/0.log" Dec 09 16:45:45 crc kubenswrapper[4894]: I1209 16:45:45.067261 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5776b569cc-b4c82_7a921fbc-ca69-439e-b357-eb15cc4de86e/barbican-worker/0.log" Dec 09 16:45:45 crc kubenswrapper[4894]: I1209 16:45:45.072138 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-5776b569cc-b4c82_7a921fbc-ca69-439e-b357-eb15cc4de86e/barbican-worker-log/0.log" Dec 09 16:45:45 crc kubenswrapper[4894]: I1209 16:45:45.278065 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c2ab328c-f7ef-432f-8e89-e7201fa7ad2c/ceilometer-central-agent/0.log" Dec 09 16:45:45 crc kubenswrapper[4894]: I1209 16:45:45.289096 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_bootstrap-edpm-deployment-openstack-edpm-ipam-sn7nv_6f620852-03c1-48ea-ae2d-b0f5de4ab16b/bootstrap-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 16:45:45 crc kubenswrapper[4894]: I1209 16:45:45.389770 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c2ab328c-f7ef-432f-8e89-e7201fa7ad2c/ceilometer-notification-agent/0.log" Dec 09 16:45:45 crc kubenswrapper[4894]: I1209 16:45:45.431522 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c2ab328c-f7ef-432f-8e89-e7201fa7ad2c/proxy-httpd/0.log" Dec 09 16:45:45 crc kubenswrapper[4894]: I1209 16:45:45.447904 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c2ab328c-f7ef-432f-8e89-e7201fa7ad2c/sg-core/0.log" Dec 09 16:45:46 crc kubenswrapper[4894]: I1209 16:45:46.382392 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-client-edpm-deployment-openstack-edpm-ipam-gjk4g_260e7406-3b55-4acf-a34a-3588a8bd1176/ceph-client-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 16:45:46 crc kubenswrapper[4894]: I1209 16:45:46.421451 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceph-hci-pre-edpm-deployment-openstack-edpm-ipam-qxckn_8bf8d33b-ff24-4ec8-9d1c-e74c9a38d952/ceph-hci-pre-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 16:45:46 crc kubenswrapper[4894]: I1209 16:45:46.691117 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_58e3314f-76b0-4e83-93ce-39c342c44acb/cinder-api-log/0.log" Dec 09 16:45:46 crc kubenswrapper[4894]: I1209 16:45:46.710105 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_58e3314f-76b0-4e83-93ce-39c342c44acb/cinder-api/0.log" Dec 09 16:45:47 crc kubenswrapper[4894]: I1209 16:45:46.988066 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_0c22d31c-b595-4f8b-bded-c20a085ab1fd/probe/0.log" Dec 09 16:45:47 crc kubenswrapper[4894]: I1209 16:45:46.999034 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-backup-0_0c22d31c-b595-4f8b-bded-c20a085ab1fd/cinder-backup/0.log" Dec 09 16:45:47 crc kubenswrapper[4894]: I1209 16:45:46.999274 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_fe06c113-787a-4daa-bba6-2c9ca0b1b4d3/cinder-scheduler/0.log" Dec 09 16:45:47 crc kubenswrapper[4894]: I1209 16:45:47.276966 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_fe06c113-787a-4daa-bba6-2c9ca0b1b4d3/probe/0.log" Dec 09 16:45:47 crc kubenswrapper[4894]: I1209 16:45:47.277324 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_73446a0b-19de-46e4-ba91-d183921ff32e/cinder-volume/0.log" Dec 09 16:45:47 crc kubenswrapper[4894]: I1209 16:45:47.289125 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-volume-volume1-0_73446a0b-19de-46e4-ba91-d183921ff32e/probe/0.log" Dec 09 16:45:47 crc kubenswrapper[4894]: I1209 16:45:47.474343 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-os-edpm-deployment-openstack-edpm-ipam-6z544_28721fcd-587f-483d-9d67-5a5fd25a1053/configure-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 16:45:47 crc kubenswrapper[4894]: I1209 16:45:47.490629 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_configure-network-edpm-deployment-openstack-edpm-ipam-vcz8l_3eebbcd6-55f4-40a5-89b3-ff5ecbddde48/configure-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 16:45:47 crc kubenswrapper[4894]: I1209 16:45:47.659452 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-67d475fdcf-c757l_28ef8474-844f-4525-8386-0e5cc06af8f7/init/0.log" Dec 09 16:45:47 crc kubenswrapper[4894]: I1209 16:45:47.904935 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_850b00b0-204b-4af9-a22d-f4264d3cb44f/glance-log/0.log" Dec 09 16:45:47 crc kubenswrapper[4894]: I1209 16:45:47.914389 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-67d475fdcf-c757l_28ef8474-844f-4525-8386-0e5cc06af8f7/dnsmasq-dns/0.log" Dec 09 16:45:47 crc kubenswrapper[4894]: I1209 16:45:47.930344 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-67d475fdcf-c757l_28ef8474-844f-4525-8386-0e5cc06af8f7/init/0.log" Dec 09 16:45:47 crc kubenswrapper[4894]: I1209 16:45:47.987745 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_850b00b0-204b-4af9-a22d-f4264d3cb44f/glance-httpd/0.log" Dec 09 16:45:48 crc kubenswrapper[4894]: I1209 16:45:48.115818 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_09510dc1-caeb-42f3-ac0f-150c7ab8365c/glance-httpd/0.log" Dec 09 16:45:48 crc kubenswrapper[4894]: I1209 16:45:48.116627 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_09510dc1-caeb-42f3-ac0f-150c7ab8365c/glance-log/0.log" Dec 09 16:45:48 crc kubenswrapper[4894]: I1209 16:45:48.277115 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6b9bcbff6b-l6n7z_6b8de55a-230d-4dbe-9f6c-aae6138d865f/horizon/0.log" Dec 09 16:45:48 crc kubenswrapper[4894]: I1209 16:45:48.447721 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizontest-tests-horizontest_6573694d-aa1d-47bc-bfcd-6d92f7416f1a/horizontest-tests-horizontest/0.log" Dec 09 16:45:48 crc kubenswrapper[4894]: I1209 16:45:48.535950 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-certs-edpm-deployment-openstack-edpm-ipam-8f2xv_f3c2da51-8857-42f2-b4a4-731927ed1599/install-certs-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 16:45:48 crc kubenswrapper[4894]: I1209 16:45:48.685147 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_install-os-edpm-deployment-openstack-edpm-ipam-vg8f5_031a7ee5-ae47-45d2-9fd2-7b8beef85242/install-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 16:45:48 crc kubenswrapper[4894]: I1209 16:45:48.902733 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-cron-29421601-2jlb4_629496e1-9961-4c1d-8f3c-9426f7ab0f72/keystone-cron/0.log" Dec 09 16:45:49 crc kubenswrapper[4894]: I1209 16:45:49.056156 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-ccc8cf6fb-9hvcd_f70e1d59-2f1f-4d88-ac43-dda028386a96/keystone-api/0.log" Dec 09 16:45:49 crc kubenswrapper[4894]: I1209 16:45:49.081905 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_62f0c278-8dcb-4e9d-8844-98f8b946edd8/kube-state-metrics/0.log" Dec 09 16:45:49 crc kubenswrapper[4894]: I1209 16:45:49.114262 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_horizon-6b9bcbff6b-l6n7z_6b8de55a-230d-4dbe-9f6c-aae6138d865f/horizon-log/0.log" Dec 09 16:45:49 crc kubenswrapper[4894]: I1209 16:45:49.163520 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_libvirt-edpm-deployment-openstack-edpm-ipam-fw9xk_f50d2ea8-df93-445e-9ce0-f5c9d4077cc1/libvirt-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 16:45:49 crc kubenswrapper[4894]: I1209 16:45:49.323882 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_289ae5f1-46ed-4165-ac04-4c3cac996270/manila-api-log/0.log" Dec 09 16:45:49 crc kubenswrapper[4894]: I1209 16:45:49.457026 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-api-0_289ae5f1-46ed-4165-ac04-4c3cac996270/manila-api/0.log" Dec 09 16:45:49 crc kubenswrapper[4894]: I1209 16:45:49.492455 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_1ba40ccc-eccc-4780-b189-f849cce60042/probe/0.log" Dec 09 16:45:49 crc kubenswrapper[4894]: I1209 16:45:49.496307 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-scheduler-0_1ba40ccc-eccc-4780-b189-f849cce60042/manila-scheduler/0.log" Dec 09 16:45:49 crc kubenswrapper[4894]: I1209 16:45:49.633864 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_45d83e58-95c2-430b-bd84-1f1a0e05d512/manila-share/0.log" Dec 09 16:45:49 crc kubenswrapper[4894]: I1209 16:45:49.676386 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_manila-share-share1-0_45d83e58-95c2-430b-bd84-1f1a0e05d512/probe/0.log" Dec 09 16:45:49 crc kubenswrapper[4894]: I1209 16:45:49.879812 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-bbb4b9577-djjjb_87c53f16-3692-4b98-9037-e2440a3d99fb/neutron-httpd/0.log" Dec 09 16:45:49 crc kubenswrapper[4894]: I1209 16:45:49.963995 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-bbb4b9577-djjjb_87c53f16-3692-4b98-9037-e2440a3d99fb/neutron-api/0.log" Dec 09 16:45:50 crc kubenswrapper[4894]: I1209 16:45:50.110736 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-metadata-edpm-deployment-openstack-edpm-ipam-nfrf5_be6e221e-e719-4805-9198-0bc9862fa4d0/neutron-metadata-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 16:45:50 crc kubenswrapper[4894]: I1209 16:45:50.399184 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_2cb294ea-866f-4569-8f83-c9567f81dfd6/nova-api-log/0.log" Dec 09 16:45:50 crc kubenswrapper[4894]: I1209 16:45:50.563473 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_e8a9b33e-85ed-4c31-abb8-21bbca492891/nova-cell0-conductor-conductor/0.log" Dec 09 16:45:50 crc kubenswrapper[4894]: I1209 16:45:50.725783 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_00ff6ff2-b452-45ae-8f9d-314363bcb126/nova-cell1-conductor-conductor/0.log" Dec 09 16:45:50 crc kubenswrapper[4894]: I1209 16:45:50.776588 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_2cb294ea-866f-4569-8f83-c9567f81dfd6/nova-api-api/0.log" Dec 09 16:45:50 crc kubenswrapper[4894]: I1209 16:45:50.965202 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_6219c187-146d-45f2-bc64-95639530b601/nova-cell1-novncproxy-novncproxy/0.log" Dec 09 16:45:51 crc kubenswrapper[4894]: I1209 16:45:51.076682 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-custom-ceph-edpm-deployment-openstack-edpm-ipam-k2r29_9625f994-137d-4957-a9f9-40bf22bf832b/nova-custom-ceph-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 16:45:51 crc kubenswrapper[4894]: E1209 16:45:51.106138 4894 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 09 16:45:51 crc kubenswrapper[4894]: I1209 16:45:51.294414 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_a1c8002b-4901-4a76-8035-66620904f04b/nova-metadata-log/0.log" Dec 09 16:45:51 crc kubenswrapper[4894]: I1209 16:45:51.472064 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_2a8b0e9c-1018-402d-9291-94588611638e/nova-scheduler-scheduler/0.log" Dec 09 16:45:51 crc kubenswrapper[4894]: I1209 16:45:51.507134 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_5caa6181-4755-426f-97fa-50a47d8ea80a/mysql-bootstrap/0.log" Dec 09 16:45:51 crc kubenswrapper[4894]: I1209 16:45:51.740969 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_5caa6181-4755-426f-97fa-50a47d8ea80a/galera/0.log" Dec 09 16:45:51 crc kubenswrapper[4894]: I1209 16:45:51.748099 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_5caa6181-4755-426f-97fa-50a47d8ea80a/mysql-bootstrap/0.log" Dec 09 16:45:51 crc kubenswrapper[4894]: I1209 16:45:51.995918 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_10d76c34-d0cc-493b-a57a-1923f8d72824/mysql-bootstrap/0.log" Dec 09 16:45:52 crc kubenswrapper[4894]: I1209 16:45:52.206187 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_10d76c34-d0cc-493b-a57a-1923f8d72824/mysql-bootstrap/0.log" Dec 09 16:45:52 crc kubenswrapper[4894]: I1209 16:45:52.342016 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_10d76c34-d0cc-493b-a57a-1923f8d72824/galera/0.log" Dec 09 16:45:52 crc kubenswrapper[4894]: I1209 16:45:52.580738 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_3752bfef-7b4f-4f5d-a449-a8b0819f6434/openstackclient/0.log" Dec 09 16:45:52 crc kubenswrapper[4894]: I1209 16:45:52.668882 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-459zv_963becb8-0d9d-48a5-b11f-29f1dd7ebb7d/openstack-network-exporter/0.log" Dec 09 16:45:52 crc kubenswrapper[4894]: I1209 16:45:52.829588 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_a1c8002b-4901-4a76-8035-66620904f04b/nova-metadata-metadata/0.log" Dec 09 16:45:52 crc kubenswrapper[4894]: I1209 16:45:52.928592 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-z7vcg_b0fedf20-9f17-4d49-bef9-c9476a5027cf/ovsdb-server-init/0.log" Dec 09 16:45:53 crc kubenswrapper[4894]: I1209 16:45:53.085972 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-z7vcg_b0fedf20-9f17-4d49-bef9-c9476a5027cf/ovs-vswitchd/0.log" Dec 09 16:45:53 crc kubenswrapper[4894]: I1209 16:45:53.101329 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-z7vcg_b0fedf20-9f17-4d49-bef9-c9476a5027cf/ovsdb-server/0.log" Dec 09 16:45:53 crc kubenswrapper[4894]: I1209 16:45:53.126125 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-z7vcg_b0fedf20-9f17-4d49-bef9-c9476a5027cf/ovsdb-server-init/0.log" Dec 09 16:45:53 crc kubenswrapper[4894]: I1209 16:45:53.538916 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c4e050ff-6a42-40ac-9298-f87ce2f7743a/openstack-network-exporter/0.log" Dec 09 16:45:53 crc kubenswrapper[4894]: I1209 16:45:53.548603 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_c4e050ff-6a42-40ac-9298-f87ce2f7743a/ovn-northd/0.log" Dec 09 16:45:53 crc kubenswrapper[4894]: I1209 16:45:53.718025 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a6e0e621-ea8e-4256-8c04-74c54686708a/openstack-network-exporter/0.log" Dec 09 16:45:53 crc kubenswrapper[4894]: I1209 16:45:53.792932 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-vk5h6_cf53d7df-bcc4-46b3-b874-dac48eca5bc4/ovn-controller/0.log" Dec 09 16:45:53 crc kubenswrapper[4894]: I1209 16:45:53.832048 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-edpm-deployment-openstack-edpm-ipam-5b5q5_9c9358fd-4505-41a1-8c05-82c59e737fa2/ovn-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 16:45:53 crc kubenswrapper[4894]: I1209 16:45:53.935172 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_a6e0e621-ea8e-4256-8c04-74c54686708a/ovsdbserver-nb/0.log" Dec 09 16:45:54 crc kubenswrapper[4894]: I1209 16:45:54.515702 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_4eebb217-538f-4f1a-a4f8-93371486803d/openstack-network-exporter/0.log" Dec 09 16:45:54 crc kubenswrapper[4894]: I1209 16:45:54.564674 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_4eebb217-538f-4f1a-a4f8-93371486803d/ovsdbserver-sb/0.log" Dec 09 16:45:54 crc kubenswrapper[4894]: I1209 16:45:54.638068 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-59f4fb4758-w8ndv_10ff6ecf-af86-4595-bff7-ad0fd49eb3e1/placement-api/0.log" Dec 09 16:45:54 crc kubenswrapper[4894]: I1209 16:45:54.820953 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-59f4fb4758-w8ndv_10ff6ecf-af86-4595-bff7-ad0fd49eb3e1/placement-log/0.log" Dec 09 16:45:54 crc kubenswrapper[4894]: I1209 16:45:54.856502 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0ad2400d-f6ca-4f83-a88f-b1bed6287315/setup-container/0.log" Dec 09 16:45:55 crc kubenswrapper[4894]: I1209 16:45:55.071961 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0ad2400d-f6ca-4f83-a88f-b1bed6287315/rabbitmq/0.log" Dec 09 16:45:55 crc kubenswrapper[4894]: I1209 16:45:55.106508 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:45:55 crc kubenswrapper[4894]: E1209 16:45:55.106812 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:45:55 crc kubenswrapper[4894]: I1209 16:45:55.181090 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_0ad2400d-f6ca-4f83-a88f-b1bed6287315/setup-container/0.log" Dec 09 16:45:55 crc kubenswrapper[4894]: I1209 16:45:55.186579 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_07e75d72-1c2a-4d08-91fb-f96409eee8b3/setup-container/0.log" Dec 09 16:45:55 crc kubenswrapper[4894]: I1209 16:45:55.286356 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_07e75d72-1c2a-4d08-91fb-f96409eee8b3/setup-container/0.log" Dec 09 16:45:55 crc kubenswrapper[4894]: I1209 16:45:55.371203 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_07e75d72-1c2a-4d08-91fb-f96409eee8b3/rabbitmq/0.log" Dec 09 16:45:55 crc kubenswrapper[4894]: I1209 16:45:55.500448 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_reboot-os-edpm-deployment-openstack-edpm-ipam-vsh6t_1b6715a4-d866-40d0-8e01-40d603a65b79/reboot-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 16:45:55 crc kubenswrapper[4894]: I1209 16:45:55.677747 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_repo-setup-edpm-deployment-openstack-edpm-ipam-vclt6_066bd81f-b24f-421c-afbc-cc34dfc33828/repo-setup-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 16:45:56 crc kubenswrapper[4894]: I1209 16:45:56.210163 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_run-os-edpm-deployment-openstack-edpm-ipam-jmhtf_c58409cd-be24-47f2-869f-6bfd530ab945/run-os-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 16:45:56 crc kubenswrapper[4894]: I1209 16:45:56.253567 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ssh-known-hosts-edpm-deployment-mxmp5_797ea3d2-8204-4940-95a7-509ab68ac7a9/ssh-known-hosts-edpm-deployment/0.log" Dec 09 16:45:56 crc kubenswrapper[4894]: I1209 16:45:56.548670 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s00-full_675a2b9e-a4fe-4ced-9997-c82358f8550c/tempest-tests-tempest-tests-runner/0.log" Dec 09 16:45:56 crc kubenswrapper[4894]: I1209 16:45:56.614888 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tempest-tests-tempest-s01-single-test_4a49e9ae-60a1-4165-ae4d-c877537d8825/tempest-tests-tempest-tests-runner/0.log" Dec 09 16:45:56 crc kubenswrapper[4894]: I1209 16:45:56.815924 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-ansibletest-ansibletest-ansibletest_b81b27e0-9170-4753-97ec-5354f447d801/test-operator-logs-container/0.log" Dec 09 16:45:56 crc kubenswrapper[4894]: I1209 16:45:56.876536 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-horizontest-horizontest-tests-horizontest_7f860d47-2922-4b88-962d-3d1d80564221/test-operator-logs-container/0.log" Dec 09 16:45:57 crc kubenswrapper[4894]: I1209 16:45:57.113379 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tempest-tempest-tests-tempest_5ecfe281-8bc0-42c3-af32-4b4b4be26a88/test-operator-logs-container/0.log" Dec 09 16:45:57 crc kubenswrapper[4894]: I1209 16:45:57.158976 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_test-operator-logs-pod-tobiko-tobiko-tests-tobiko_8135eb77-f9c3-4570-b4d8-c7d87f8cdf52/test-operator-logs-container/0.log" Dec 09 16:45:57 crc kubenswrapper[4894]: I1209 16:45:57.415023 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tobiko-tests-tobiko-s00-podified-functional_f3302421-5f23-4e31-b6b7-9ea816c8fdea/tobiko-tests-tobiko/0.log" Dec 09 16:45:57 crc kubenswrapper[4894]: I1209 16:45:57.451521 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_tobiko-tests-tobiko-s01-sanity_3dd4c51e-1b24-48e9-a722-80c11a2460fd/tobiko-tests-tobiko/0.log" Dec 09 16:45:57 crc kubenswrapper[4894]: I1209 16:45:57.899899 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_validate-network-edpm-deployment-openstack-edpm-ipam-tmmsg_ac0dac47-1921-41cf-bc20-c0259c1f75d8/validate-network-edpm-deployment-openstack-edpm-ipam/0.log" Dec 09 16:45:58 crc kubenswrapper[4894]: I1209 16:45:58.754790 4894 scope.go:117] "RemoveContainer" containerID="8529f745f32501dce0e99d50fb124aa049cd3448ed68dd213d19bec97dcd1c85" Dec 09 16:46:06 crc kubenswrapper[4894]: I1209 16:46:06.134857 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:46:06 crc kubenswrapper[4894]: E1209 16:46:06.135842 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:46:09 crc kubenswrapper[4894]: I1209 16:46:09.612259 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_9ef2f4b3-b303-4b17-b4e8-fc2b8929bdfb/memcached/0.log" Dec 09 16:46:17 crc kubenswrapper[4894]: I1209 16:46:17.107239 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:46:17 crc kubenswrapper[4894]: E1209 16:46:17.107958 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:46:27 crc kubenswrapper[4894]: I1209 16:46:27.741532 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b_caf14a3e-1621-42fc-9f4f-95d64b325062/util/0.log" Dec 09 16:46:27 crc kubenswrapper[4894]: I1209 16:46:27.890583 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b_caf14a3e-1621-42fc-9f4f-95d64b325062/util/0.log" Dec 09 16:46:28 crc kubenswrapper[4894]: I1209 16:46:28.106599 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:46:28 crc kubenswrapper[4894]: E1209 16:46:28.106965 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:46:28 crc kubenswrapper[4894]: I1209 16:46:28.119612 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b_caf14a3e-1621-42fc-9f4f-95d64b325062/pull/0.log" Dec 09 16:46:28 crc kubenswrapper[4894]: I1209 16:46:28.119860 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b_caf14a3e-1621-42fc-9f4f-95d64b325062/pull/0.log" Dec 09 16:46:28 crc kubenswrapper[4894]: I1209 16:46:28.119951 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b_caf14a3e-1621-42fc-9f4f-95d64b325062/pull/0.log" Dec 09 16:46:28 crc kubenswrapper[4894]: I1209 16:46:28.272744 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b_caf14a3e-1621-42fc-9f4f-95d64b325062/util/0.log" Dec 09 16:46:28 crc kubenswrapper[4894]: I1209 16:46:28.312061 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_a202a8d0d29caf7438df631524a4272602a9619288fa076173728c3fc3cp65b_caf14a3e-1621-42fc-9f4f-95d64b325062/extract/0.log" Dec 09 16:46:28 crc kubenswrapper[4894]: I1209 16:46:28.349478 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-xbxkp_52b39bc2-f99f-4865-8fb2-d4377ef35e88/kube-rbac-proxy/0.log" Dec 09 16:46:28 crc kubenswrapper[4894]: I1209 16:46:28.509004 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-7d9dfd778-xbxkp_52b39bc2-f99f-4865-8fb2-d4377ef35e88/manager/0.log" Dec 09 16:46:28 crc kubenswrapper[4894]: I1209 16:46:28.574257 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-rjhh2_cfbfae03-83af-4dd6-b04b-6e094638d96e/kube-rbac-proxy/0.log" Dec 09 16:46:28 crc kubenswrapper[4894]: I1209 16:46:28.670750 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-6c677c69b-rjhh2_cfbfae03-83af-4dd6-b04b-6e094638d96e/manager/0.log" Dec 09 16:46:28 crc kubenswrapper[4894]: I1209 16:46:28.788907 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-w47lh_7b9d28bb-988b-4b71-99bb-df62f57debc8/kube-rbac-proxy/0.log" Dec 09 16:46:28 crc kubenswrapper[4894]: I1209 16:46:28.792236 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-697fb699cf-w47lh_7b9d28bb-988b-4b71-99bb-df62f57debc8/manager/0.log" Dec 09 16:46:29 crc kubenswrapper[4894]: I1209 16:46:29.008494 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-nv5mv_811e1bfc-20a8-4268-9bda-88dd4e1b33ad/kube-rbac-proxy/0.log" Dec 09 16:46:29 crc kubenswrapper[4894]: I1209 16:46:29.051958 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-5697bb5779-nv5mv_811e1bfc-20a8-4268-9bda-88dd4e1b33ad/manager/0.log" Dec 09 16:46:29 crc kubenswrapper[4894]: I1209 16:46:29.139531 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-v2jx9_e2902997-82a3-4e80-b50b-6fbc3f290e61/kube-rbac-proxy/0.log" Dec 09 16:46:29 crc kubenswrapper[4894]: I1209 16:46:29.223356 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-5f64f6f8bb-v2jx9_e2902997-82a3-4e80-b50b-6fbc3f290e61/manager/0.log" Dec 09 16:46:29 crc kubenswrapper[4894]: I1209 16:46:29.228209 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-6d2gw_63f9a2a8-9f28-4ee7-a564-9c449f15917b/kube-rbac-proxy/0.log" Dec 09 16:46:29 crc kubenswrapper[4894]: I1209 16:46:29.353182 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-68c6d99b8f-6d2gw_63f9a2a8-9f28-4ee7-a564-9c449f15917b/manager/0.log" Dec 09 16:46:29 crc kubenswrapper[4894]: I1209 16:46:29.407792 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-mxngk_622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7/kube-rbac-proxy/0.log" Dec 09 16:46:29 crc kubenswrapper[4894]: I1209 16:46:29.634514 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-4qzlb_7dd34c7e-4139-4dce-873d-32ed62e502ae/kube-rbac-proxy/0.log" Dec 09 16:46:29 crc kubenswrapper[4894]: I1209 16:46:29.652524 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-967d97867-4qzlb_7dd34c7e-4139-4dce-873d-32ed62e502ae/manager/0.log" Dec 09 16:46:29 crc kubenswrapper[4894]: I1209 16:46:29.690937 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-78d48bff9d-mxngk_622a8fae-c8e4-4ec9-bd3f-edbb1e2a5ed7/manager/0.log" Dec 09 16:46:29 crc kubenswrapper[4894]: I1209 16:46:29.897108 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-5kgk4_4aa1a97a-d5a6-42e1-9db7-94ca1cad21ff/manager/0.log" Dec 09 16:46:29 crc kubenswrapper[4894]: I1209 16:46:29.920575 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-7765d96ddf-5kgk4_4aa1a97a-d5a6-42e1-9db7-94ca1cad21ff/kube-rbac-proxy/0.log" Dec 09 16:46:30 crc kubenswrapper[4894]: I1209 16:46:30.061430 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-5ptth_d85fd4c5-d08d-4bda-b8b9-aad2165e930c/kube-rbac-proxy/0.log" Dec 09 16:46:30 crc kubenswrapper[4894]: I1209 16:46:30.168748 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-5b5fd79c9c-5ptth_d85fd4c5-d08d-4bda-b8b9-aad2165e930c/manager/0.log" Dec 09 16:46:30 crc kubenswrapper[4894]: I1209 16:46:30.183937 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-9wtvr_ab257a96-149b-4abf-b4b1-8d1beb8e873d/kube-rbac-proxy/0.log" Dec 09 16:46:30 crc kubenswrapper[4894]: I1209 16:46:30.320059 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-79c8c4686c-9wtvr_ab257a96-149b-4abf-b4b1-8d1beb8e873d/manager/0.log" Dec 09 16:46:30 crc kubenswrapper[4894]: I1209 16:46:30.405577 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-b966c_1fb9fefc-7b00-4116-abc7-6f8bdf847f82/kube-rbac-proxy/0.log" Dec 09 16:46:30 crc kubenswrapper[4894]: I1209 16:46:30.438459 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-5fdfd5b6b5-b966c_1fb9fefc-7b00-4116-abc7-6f8bdf847f82/manager/0.log" Dec 09 16:46:30 crc kubenswrapper[4894]: I1209 16:46:30.525944 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-7qv58_e177f09b-5191-4e61-af98-91aa6bf25fc3/kube-rbac-proxy/0.log" Dec 09 16:46:30 crc kubenswrapper[4894]: I1209 16:46:30.727500 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-mhx7f_cfcfd730-cad4-4774-91fe-7c3d7692f011/kube-rbac-proxy/0.log" Dec 09 16:46:30 crc kubenswrapper[4894]: I1209 16:46:30.816929 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-998648c74-mhx7f_cfcfd730-cad4-4774-91fe-7c3d7692f011/manager/0.log" Dec 09 16:46:30 crc kubenswrapper[4894]: I1209 16:46:30.820656 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-697bc559fc-7qv58_e177f09b-5191-4e61-af98-91aa6bf25fc3/manager/0.log" Dec 09 16:46:30 crc kubenswrapper[4894]: I1209 16:46:30.951914 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-694d6cfbd667tpw_9557c24e-8469-46ec-9bab-04e337606beb/kube-rbac-proxy/0.log" Dec 09 16:46:31 crc kubenswrapper[4894]: I1209 16:46:31.003250 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-694d6cfbd667tpw_9557c24e-8469-46ec-9bab-04e337606beb/manager/0.log" Dec 09 16:46:31 crc kubenswrapper[4894]: I1209 16:46:31.251358 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-c4486db64-xg2r2_6a5bd235-8edd-4264-85f0-fabe63b56012/operator/0.log" Dec 09 16:46:31 crc kubenswrapper[4894]: I1209 16:46:31.350934 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-m8kvv_cfb263e2-5739-43ed-965a-575f4cb83652/registry-server/0.log" Dec 09 16:46:31 crc kubenswrapper[4894]: I1209 16:46:31.451218 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-tdcqr_4ca57622-56aa-4588-a1bd-f28972322b58/kube-rbac-proxy/0.log" Dec 09 16:46:31 crc kubenswrapper[4894]: I1209 16:46:31.592404 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-b6456fdb6-tdcqr_4ca57622-56aa-4588-a1bd-f28972322b58/manager/0.log" Dec 09 16:46:31 crc kubenswrapper[4894]: I1209 16:46:31.765360 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-wpj5r_faf8439c-1fc7-4316-8b58-31c5a892258b/kube-rbac-proxy/0.log" Dec 09 16:46:31 crc kubenswrapper[4894]: I1209 16:46:31.928240 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-78f8948974-wpj5r_faf8439c-1fc7-4316-8b58-31c5a892258b/manager/0.log" Dec 09 16:46:32 crc kubenswrapper[4894]: I1209 16:46:32.001288 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-668c99d594-6fdzd_7700fe8f-1bda-4f92-81ff-04463e0e4fdf/operator/0.log" Dec 09 16:46:32 crc kubenswrapper[4894]: I1209 16:46:32.678901 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-668858c49-xgc5f_cc2f225f-a9cd-493a-bb6b-6d8adb308080/manager/0.log" Dec 09 16:46:32 crc kubenswrapper[4894]: I1209 16:46:32.805739 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-wxcz2_66650287-6d2b-412e-8b94-fcfedd5a6194/manager/0.log" Dec 09 16:46:32 crc kubenswrapper[4894]: I1209 16:46:32.846560 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-frzr9_7f99fc36-cb37-456c-bb67-5b4f27d00b80/kube-rbac-proxy/0.log" Dec 09 16:46:32 crc kubenswrapper[4894]: I1209 16:46:32.859418 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-9d58d64bc-wxcz2_66650287-6d2b-412e-8b94-fcfedd5a6194/kube-rbac-proxy/0.log" Dec 09 16:46:32 crc kubenswrapper[4894]: I1209 16:46:32.992747 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-58d5ff84df-frzr9_7f99fc36-cb37-456c-bb67-5b4f27d00b80/manager/0.log" Dec 09 16:46:33 crc kubenswrapper[4894]: I1209 16:46:33.121556 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-6dc759c869-b8lkg_23370728-b91c-4d67-9c3b-a61aef727c25/kube-rbac-proxy/0.log" Dec 09 16:46:33 crc kubenswrapper[4894]: I1209 16:46:33.141850 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-6dc759c869-b8lkg_23370728-b91c-4d67-9c3b-a61aef727c25/manager/0.log" Dec 09 16:46:33 crc kubenswrapper[4894]: I1209 16:46:33.214286 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-c9v7t_29ac0821-c99d-4b02-8ea7-9985e535305c/kube-rbac-proxy/0.log" Dec 09 16:46:33 crc kubenswrapper[4894]: I1209 16:46:33.302142 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-667bd8d554-c9v7t_29ac0821-c99d-4b02-8ea7-9985e535305c/manager/0.log" Dec 09 16:46:41 crc kubenswrapper[4894]: I1209 16:46:41.106935 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:46:41 crc kubenswrapper[4894]: E1209 16:46:41.107752 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:46:52 crc kubenswrapper[4894]: I1209 16:46:52.106913 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:46:52 crc kubenswrapper[4894]: E1209 16:46:52.107861 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:46:53 crc kubenswrapper[4894]: E1209 16:46:53.106208 4894 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 09 16:46:54 crc kubenswrapper[4894]: I1209 16:46:54.410798 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-9qzp8_9c59e082-e6c9-4e92-8f39-e6f11f2d4941/control-plane-machine-set-operator/0.log" Dec 09 16:46:54 crc kubenswrapper[4894]: I1209 16:46:54.428932 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-rvzr9_66f45f17-addc-4737-a47c-1ab34a289659/kube-rbac-proxy/0.log" Dec 09 16:46:54 crc kubenswrapper[4894]: I1209 16:46:54.670033 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-rvzr9_66f45f17-addc-4737-a47c-1ab34a289659/machine-api-operator/0.log" Dec 09 16:47:05 crc kubenswrapper[4894]: I1209 16:47:05.106301 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:47:05 crc kubenswrapper[4894]: E1209 16:47:05.108354 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:47:12 crc kubenswrapper[4894]: I1209 16:47:12.919144 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-kwc95_74585f72-f240-4bdd-8f23-01874e9551fe/cert-manager-controller/0.log" Dec 09 16:47:13 crc kubenswrapper[4894]: I1209 16:47:13.192820 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-2f27v_d7727746-e66c-4645-af8d-56efc6b4ed82/cert-manager-cainjector/0.log" Dec 09 16:47:13 crc kubenswrapper[4894]: I1209 16:47:13.236412 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-k8tfx_9881dac0-8703-4808-8e2d-45d2df155dcf/cert-manager-webhook/0.log" Dec 09 16:47:20 crc kubenswrapper[4894]: I1209 16:47:20.106642 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:47:20 crc kubenswrapper[4894]: E1209 16:47:20.107727 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:47:26 crc kubenswrapper[4894]: I1209 16:47:26.639175 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-7fbb5f6569-8vs59_61dcd6dd-61fc-4b15-8c34-b7b27775f1cd/nmstate-console-plugin/0.log" Dec 09 16:47:27 crc kubenswrapper[4894]: I1209 16:47:27.278195 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-tclzs_81fe83b5-37f9-499e-af80-602c302d905c/kube-rbac-proxy/0.log" Dec 09 16:47:27 crc kubenswrapper[4894]: I1209 16:47:27.295997 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-4vf4p_34e69c07-d85c-4a34-afe7-287647f01876/nmstate-handler/0.log" Dec 09 16:47:27 crc kubenswrapper[4894]: I1209 16:47:27.311832 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-7f946cbc9-tclzs_81fe83b5-37f9-499e-af80-602c302d905c/nmstate-metrics/0.log" Dec 09 16:47:27 crc kubenswrapper[4894]: I1209 16:47:27.433259 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-5b5b58f5c8-5ctgb_a6aa74db-0590-4c30-ae7a-164d405f3944/nmstate-operator/0.log" Dec 09 16:47:27 crc kubenswrapper[4894]: I1209 16:47:27.585670 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-5f6d4c5ccb-5zmgn_2cbfef40-9c05-4039-b1db-d47e119ab8da/nmstate-webhook/0.log" Dec 09 16:47:35 crc kubenswrapper[4894]: I1209 16:47:35.106547 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:47:35 crc kubenswrapper[4894]: E1209 16:47:35.107413 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:47:43 crc kubenswrapper[4894]: I1209 16:47:43.190808 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-tjjzx_26ecc2b2-2443-4069-bfda-934ff3835cf8/kube-rbac-proxy/0.log" Dec 09 16:47:43 crc kubenswrapper[4894]: I1209 16:47:43.321771 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-f8648f98b-tjjzx_26ecc2b2-2443-4069-bfda-934ff3835cf8/controller/0.log" Dec 09 16:47:43 crc kubenswrapper[4894]: I1209 16:47:43.362775 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jdblb_2b7c2f69-58e9-4aad-9900-22befc1bc6c6/cp-frr-files/0.log" Dec 09 16:47:43 crc kubenswrapper[4894]: I1209 16:47:43.517107 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jdblb_2b7c2f69-58e9-4aad-9900-22befc1bc6c6/cp-reloader/0.log" Dec 09 16:47:43 crc kubenswrapper[4894]: I1209 16:47:43.518706 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jdblb_2b7c2f69-58e9-4aad-9900-22befc1bc6c6/cp-frr-files/0.log" Dec 09 16:47:43 crc kubenswrapper[4894]: I1209 16:47:43.569592 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jdblb_2b7c2f69-58e9-4aad-9900-22befc1bc6c6/cp-metrics/0.log" Dec 09 16:47:43 crc kubenswrapper[4894]: I1209 16:47:43.599060 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jdblb_2b7c2f69-58e9-4aad-9900-22befc1bc6c6/cp-reloader/0.log" Dec 09 16:47:43 crc kubenswrapper[4894]: I1209 16:47:43.802121 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jdblb_2b7c2f69-58e9-4aad-9900-22befc1bc6c6/cp-frr-files/0.log" Dec 09 16:47:43 crc kubenswrapper[4894]: I1209 16:47:43.809107 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jdblb_2b7c2f69-58e9-4aad-9900-22befc1bc6c6/cp-metrics/0.log" Dec 09 16:47:43 crc kubenswrapper[4894]: I1209 16:47:43.811852 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jdblb_2b7c2f69-58e9-4aad-9900-22befc1bc6c6/cp-reloader/0.log" Dec 09 16:47:43 crc kubenswrapper[4894]: I1209 16:47:43.817087 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jdblb_2b7c2f69-58e9-4aad-9900-22befc1bc6c6/cp-metrics/0.log" Dec 09 16:47:43 crc kubenswrapper[4894]: I1209 16:47:43.985027 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jdblb_2b7c2f69-58e9-4aad-9900-22befc1bc6c6/cp-reloader/0.log" Dec 09 16:47:44 crc kubenswrapper[4894]: I1209 16:47:44.010705 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jdblb_2b7c2f69-58e9-4aad-9900-22befc1bc6c6/cp-frr-files/0.log" Dec 09 16:47:44 crc kubenswrapper[4894]: I1209 16:47:44.019464 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jdblb_2b7c2f69-58e9-4aad-9900-22befc1bc6c6/cp-metrics/0.log" Dec 09 16:47:44 crc kubenswrapper[4894]: I1209 16:47:44.020112 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jdblb_2b7c2f69-58e9-4aad-9900-22befc1bc6c6/controller/0.log" Dec 09 16:47:44 crc kubenswrapper[4894]: I1209 16:47:44.175631 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jdblb_2b7c2f69-58e9-4aad-9900-22befc1bc6c6/kube-rbac-proxy/0.log" Dec 09 16:47:44 crc kubenswrapper[4894]: I1209 16:47:44.204182 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jdblb_2b7c2f69-58e9-4aad-9900-22befc1bc6c6/frr-metrics/0.log" Dec 09 16:47:44 crc kubenswrapper[4894]: I1209 16:47:44.273189 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jdblb_2b7c2f69-58e9-4aad-9900-22befc1bc6c6/kube-rbac-proxy-frr/0.log" Dec 09 16:47:44 crc kubenswrapper[4894]: I1209 16:47:44.441859 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jdblb_2b7c2f69-58e9-4aad-9900-22befc1bc6c6/reloader/0.log" Dec 09 16:47:44 crc kubenswrapper[4894]: I1209 16:47:44.513495 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-7fcb986d4-5dr97_ba40556a-56db-491c-b0b0-e7b7f8ec1b6a/frr-k8s-webhook-server/0.log" Dec 09 16:47:44 crc kubenswrapper[4894]: I1209 16:47:44.634028 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-85ff77c5c9-mzmjf_2d23be67-e833-46bd-a293-a41fc8144ca1/manager/0.log" Dec 09 16:47:44 crc kubenswrapper[4894]: I1209 16:47:44.883967 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-59cdbccb6-v6f5s_587142f6-a6a8-4500-97d5-82f0d5fc04c7/webhook-server/0.log" Dec 09 16:47:44 crc kubenswrapper[4894]: I1209 16:47:44.981585 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8zntm_d0f8c048-3892-4c12-a074-b7c4c5e85b87/kube-rbac-proxy/0.log" Dec 09 16:47:45 crc kubenswrapper[4894]: I1209 16:47:45.856747 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-8zntm_d0f8c048-3892-4c12-a074-b7c4c5e85b87/speaker/0.log" Dec 09 16:47:46 crc kubenswrapper[4894]: I1209 16:47:46.257511 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-jdblb_2b7c2f69-58e9-4aad-9900-22befc1bc6c6/frr/0.log" Dec 09 16:47:50 crc kubenswrapper[4894]: I1209 16:47:50.107584 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:47:50 crc kubenswrapper[4894]: E1209 16:47:50.108826 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:47:59 crc kubenswrapper[4894]: I1209 16:47:59.243091 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq_f961acc2-dd50-4607-8ff5-d06a51bddf97/util/0.log" Dec 09 16:47:59 crc kubenswrapper[4894]: I1209 16:47:59.429382 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq_f961acc2-dd50-4607-8ff5-d06a51bddf97/pull/0.log" Dec 09 16:47:59 crc kubenswrapper[4894]: I1209 16:47:59.431247 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq_f961acc2-dd50-4607-8ff5-d06a51bddf97/util/0.log" Dec 09 16:47:59 crc kubenswrapper[4894]: I1209 16:47:59.475339 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq_f961acc2-dd50-4607-8ff5-d06a51bddf97/pull/0.log" Dec 09 16:47:59 crc kubenswrapper[4894]: I1209 16:47:59.634334 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq_f961acc2-dd50-4607-8ff5-d06a51bddf97/util/0.log" Dec 09 16:47:59 crc kubenswrapper[4894]: I1209 16:47:59.642051 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq_f961acc2-dd50-4607-8ff5-d06a51bddf97/pull/0.log" Dec 09 16:47:59 crc kubenswrapper[4894]: I1209 16:47:59.659333 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_5064f9f8917b246f69f5d7fc025e7e6c34236c02bca31167615d38212fcq6sq_f961acc2-dd50-4607-8ff5-d06a51bddf97/extract/0.log" Dec 09 16:47:59 crc kubenswrapper[4894]: I1209 16:47:59.824387 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl_423c01ef-021c-44a3-a606-df1bfef275df/util/0.log" Dec 09 16:47:59 crc kubenswrapper[4894]: I1209 16:47:59.968570 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl_423c01ef-021c-44a3-a606-df1bfef275df/util/0.log" Dec 09 16:47:59 crc kubenswrapper[4894]: I1209 16:47:59.972796 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl_423c01ef-021c-44a3-a606-df1bfef275df/pull/0.log" Dec 09 16:47:59 crc kubenswrapper[4894]: I1209 16:47:59.992703 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl_423c01ef-021c-44a3-a606-df1bfef275df/pull/0.log" Dec 09 16:48:00 crc kubenswrapper[4894]: I1209 16:48:00.211907 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl_423c01ef-021c-44a3-a606-df1bfef275df/pull/0.log" Dec 09 16:48:00 crc kubenswrapper[4894]: I1209 16:48:00.235121 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl_423c01ef-021c-44a3-a606-df1bfef275df/util/0.log" Dec 09 16:48:00 crc kubenswrapper[4894]: I1209 16:48:00.286611 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_af69698b82fdf008f5ff9e195cf808a654240e16b13dcd924b74994f83lmpjl_423c01ef-021c-44a3-a606-df1bfef275df/extract/0.log" Dec 09 16:48:00 crc kubenswrapper[4894]: I1209 16:48:00.399004 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lnlst_74ee2091-d89d-4945-af03-5b9887bdfa24/extract-utilities/0.log" Dec 09 16:48:00 crc kubenswrapper[4894]: I1209 16:48:00.581805 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lnlst_74ee2091-d89d-4945-af03-5b9887bdfa24/extract-content/0.log" Dec 09 16:48:00 crc kubenswrapper[4894]: I1209 16:48:00.601403 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lnlst_74ee2091-d89d-4945-af03-5b9887bdfa24/extract-utilities/0.log" Dec 09 16:48:00 crc kubenswrapper[4894]: I1209 16:48:00.613926 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lnlst_74ee2091-d89d-4945-af03-5b9887bdfa24/extract-content/0.log" Dec 09 16:48:00 crc kubenswrapper[4894]: I1209 16:48:00.749353 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lnlst_74ee2091-d89d-4945-af03-5b9887bdfa24/extract-utilities/0.log" Dec 09 16:48:00 crc kubenswrapper[4894]: I1209 16:48:00.766468 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lnlst_74ee2091-d89d-4945-af03-5b9887bdfa24/extract-content/0.log" Dec 09 16:48:00 crc kubenswrapper[4894]: I1209 16:48:00.944364 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rb4sl_b469772b-9a9b-43dd-92ea-059528977c3a/extract-utilities/0.log" Dec 09 16:48:01 crc kubenswrapper[4894]: I1209 16:48:01.046947 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-lnlst_74ee2091-d89d-4945-af03-5b9887bdfa24/registry-server/0.log" Dec 09 16:48:01 crc kubenswrapper[4894]: I1209 16:48:01.183341 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rb4sl_b469772b-9a9b-43dd-92ea-059528977c3a/extract-utilities/0.log" Dec 09 16:48:01 crc kubenswrapper[4894]: I1209 16:48:01.187610 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rb4sl_b469772b-9a9b-43dd-92ea-059528977c3a/extract-content/0.log" Dec 09 16:48:01 crc kubenswrapper[4894]: I1209 16:48:01.198674 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rb4sl_b469772b-9a9b-43dd-92ea-059528977c3a/extract-content/0.log" Dec 09 16:48:01 crc kubenswrapper[4894]: I1209 16:48:01.335832 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rb4sl_b469772b-9a9b-43dd-92ea-059528977c3a/extract-utilities/0.log" Dec 09 16:48:01 crc kubenswrapper[4894]: I1209 16:48:01.390905 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rb4sl_b469772b-9a9b-43dd-92ea-059528977c3a/extract-content/0.log" Dec 09 16:48:01 crc kubenswrapper[4894]: I1209 16:48:01.522291 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-v5dgx_02757ffa-44f3-4f5d-ad50-7d7044ce3a24/marketplace-operator/0.log" Dec 09 16:48:03 crc kubenswrapper[4894]: I1209 16:48:03.107975 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:48:03 crc kubenswrapper[4894]: E1209 16:48:03.108469 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:48:03 crc kubenswrapper[4894]: I1209 16:48:03.140746 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zg6pn_992e28b8-b2cb-4795-b27e-76b87af880c3/extract-utilities/0.log" Dec 09 16:48:03 crc kubenswrapper[4894]: I1209 16:48:03.499423 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zg6pn_992e28b8-b2cb-4795-b27e-76b87af880c3/extract-content/0.log" Dec 09 16:48:03 crc kubenswrapper[4894]: I1209 16:48:03.516753 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zg6pn_992e28b8-b2cb-4795-b27e-76b87af880c3/extract-utilities/0.log" Dec 09 16:48:03 crc kubenswrapper[4894]: I1209 16:48:03.572097 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zg6pn_992e28b8-b2cb-4795-b27e-76b87af880c3/extract-content/0.log" Dec 09 16:48:03 crc kubenswrapper[4894]: I1209 16:48:03.756115 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zg6pn_992e28b8-b2cb-4795-b27e-76b87af880c3/extract-utilities/0.log" Dec 09 16:48:03 crc kubenswrapper[4894]: I1209 16:48:03.785913 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zg6pn_992e28b8-b2cb-4795-b27e-76b87af880c3/extract-content/0.log" Dec 09 16:48:03 crc kubenswrapper[4894]: I1209 16:48:03.869159 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-rb4sl_b469772b-9a9b-43dd-92ea-059528977c3a/registry-server/0.log" Dec 09 16:48:03 crc kubenswrapper[4894]: I1209 16:48:03.955065 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-88kvs_da62df93-33dd-4f5c-85f6-e82f8b1f1b2a/extract-utilities/0.log" Dec 09 16:48:04 crc kubenswrapper[4894]: I1209 16:48:04.018728 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-zg6pn_992e28b8-b2cb-4795-b27e-76b87af880c3/registry-server/0.log" Dec 09 16:48:04 crc kubenswrapper[4894]: I1209 16:48:04.168280 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-88kvs_da62df93-33dd-4f5c-85f6-e82f8b1f1b2a/extract-utilities/0.log" Dec 09 16:48:04 crc kubenswrapper[4894]: I1209 16:48:04.211258 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-88kvs_da62df93-33dd-4f5c-85f6-e82f8b1f1b2a/extract-content/0.log" Dec 09 16:48:04 crc kubenswrapper[4894]: I1209 16:48:04.224945 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-88kvs_da62df93-33dd-4f5c-85f6-e82f8b1f1b2a/extract-content/0.log" Dec 09 16:48:04 crc kubenswrapper[4894]: I1209 16:48:04.369417 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-88kvs_da62df93-33dd-4f5c-85f6-e82f8b1f1b2a/extract-content/0.log" Dec 09 16:48:04 crc kubenswrapper[4894]: I1209 16:48:04.398452 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-88kvs_da62df93-33dd-4f5c-85f6-e82f8b1f1b2a/extract-utilities/0.log" Dec 09 16:48:04 crc kubenswrapper[4894]: I1209 16:48:04.534902 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-88kvs_da62df93-33dd-4f5c-85f6-e82f8b1f1b2a/registry-server/0.log" Dec 09 16:48:14 crc kubenswrapper[4894]: I1209 16:48:14.106745 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:48:14 crc kubenswrapper[4894]: E1209 16:48:14.107466 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:48:22 crc kubenswrapper[4894]: E1209 16:48:22.107148 4894 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 09 16:48:25 crc kubenswrapper[4894]: I1209 16:48:25.107138 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:48:25 crc kubenswrapper[4894]: E1209 16:48:25.107974 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:48:36 crc kubenswrapper[4894]: I1209 16:48:36.114037 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:48:36 crc kubenswrapper[4894]: E1209 16:48:36.114856 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:48:49 crc kubenswrapper[4894]: I1209 16:48:49.107296 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:48:49 crc kubenswrapper[4894]: E1209 16:48:49.108380 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:49:02 crc kubenswrapper[4894]: I1209 16:49:02.109193 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:49:02 crc kubenswrapper[4894]: E1209 16:49:02.110029 4894 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-rj7cb_openshift-machine-config-operator(cfa97fcb-7e7c-4202-a2d0-d283fa49b112)\"" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" Dec 09 16:49:17 crc kubenswrapper[4894]: I1209 16:49:17.106704 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:49:18 crc kubenswrapper[4894]: I1209 16:49:18.360670 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerStarted","Data":"d6834fb5fc2cfe31b11cb42928fcf576c01fe5393bc29ae78a328cdf98460675"} Dec 09 16:49:43 crc kubenswrapper[4894]: E1209 16:49:43.106681 4894 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 09 16:49:54 crc kubenswrapper[4894]: I1209 16:49:54.743687 4894 generic.go:334] "Generic (PLEG): container finished" podID="803d496d-7f84-4100-9f68-e2413ff22a54" containerID="4d585bb631c76873f173ca309f0f4ef760b09bc83ba0e5ad2fdda994558fd24b" exitCode=0 Dec 09 16:49:54 crc kubenswrapper[4894]: I1209 16:49:54.743792 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-pn9dz/must-gather-xjnpp" event={"ID":"803d496d-7f84-4100-9f68-e2413ff22a54","Type":"ContainerDied","Data":"4d585bb631c76873f173ca309f0f4ef760b09bc83ba0e5ad2fdda994558fd24b"} Dec 09 16:49:54 crc kubenswrapper[4894]: I1209 16:49:54.744721 4894 scope.go:117] "RemoveContainer" containerID="4d585bb631c76873f173ca309f0f4ef760b09bc83ba0e5ad2fdda994558fd24b" Dec 09 16:49:55 crc kubenswrapper[4894]: I1209 16:49:55.314100 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pn9dz_must-gather-xjnpp_803d496d-7f84-4100-9f68-e2413ff22a54/gather/0.log" Dec 09 16:50:03 crc kubenswrapper[4894]: I1209 16:50:03.584909 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-pn9dz/must-gather-xjnpp"] Dec 09 16:50:03 crc kubenswrapper[4894]: I1209 16:50:03.585905 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-pn9dz/must-gather-xjnpp" podUID="803d496d-7f84-4100-9f68-e2413ff22a54" containerName="copy" containerID="cri-o://ec73518edbce09c541073097d6977020ce9e53688c23c5782022b8bcf6012a9a" gracePeriod=2 Dec 09 16:50:03 crc kubenswrapper[4894]: I1209 16:50:03.594325 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-pn9dz/must-gather-xjnpp"] Dec 09 16:50:03 crc kubenswrapper[4894]: I1209 16:50:03.840845 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pn9dz_must-gather-xjnpp_803d496d-7f84-4100-9f68-e2413ff22a54/copy/0.log" Dec 09 16:50:03 crc kubenswrapper[4894]: I1209 16:50:03.841676 4894 generic.go:334] "Generic (PLEG): container finished" podID="803d496d-7f84-4100-9f68-e2413ff22a54" containerID="ec73518edbce09c541073097d6977020ce9e53688c23c5782022b8bcf6012a9a" exitCode=143 Dec 09 16:50:04 crc kubenswrapper[4894]: I1209 16:50:04.200886 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pn9dz_must-gather-xjnpp_803d496d-7f84-4100-9f68-e2413ff22a54/copy/0.log" Dec 09 16:50:04 crc kubenswrapper[4894]: I1209 16:50:04.201363 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pn9dz/must-gather-xjnpp" Dec 09 16:50:04 crc kubenswrapper[4894]: I1209 16:50:04.232108 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/803d496d-7f84-4100-9f68-e2413ff22a54-must-gather-output\") pod \"803d496d-7f84-4100-9f68-e2413ff22a54\" (UID: \"803d496d-7f84-4100-9f68-e2413ff22a54\") " Dec 09 16:50:04 crc kubenswrapper[4894]: I1209 16:50:04.232201 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ft5f4\" (UniqueName: \"kubernetes.io/projected/803d496d-7f84-4100-9f68-e2413ff22a54-kube-api-access-ft5f4\") pod \"803d496d-7f84-4100-9f68-e2413ff22a54\" (UID: \"803d496d-7f84-4100-9f68-e2413ff22a54\") " Dec 09 16:50:04 crc kubenswrapper[4894]: I1209 16:50:04.247270 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/803d496d-7f84-4100-9f68-e2413ff22a54-kube-api-access-ft5f4" (OuterVolumeSpecName: "kube-api-access-ft5f4") pod "803d496d-7f84-4100-9f68-e2413ff22a54" (UID: "803d496d-7f84-4100-9f68-e2413ff22a54"). InnerVolumeSpecName "kube-api-access-ft5f4". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:50:04 crc kubenswrapper[4894]: I1209 16:50:04.335138 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ft5f4\" (UniqueName: \"kubernetes.io/projected/803d496d-7f84-4100-9f68-e2413ff22a54-kube-api-access-ft5f4\") on node \"crc\" DevicePath \"\"" Dec 09 16:50:04 crc kubenswrapper[4894]: I1209 16:50:04.373142 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/803d496d-7f84-4100-9f68-e2413ff22a54-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "803d496d-7f84-4100-9f68-e2413ff22a54" (UID: "803d496d-7f84-4100-9f68-e2413ff22a54"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:50:04 crc kubenswrapper[4894]: I1209 16:50:04.437857 4894 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/803d496d-7f84-4100-9f68-e2413ff22a54-must-gather-output\") on node \"crc\" DevicePath \"\"" Dec 09 16:50:04 crc kubenswrapper[4894]: I1209 16:50:04.852763 4894 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-pn9dz_must-gather-xjnpp_803d496d-7f84-4100-9f68-e2413ff22a54/copy/0.log" Dec 09 16:50:04 crc kubenswrapper[4894]: I1209 16:50:04.853265 4894 scope.go:117] "RemoveContainer" containerID="ec73518edbce09c541073097d6977020ce9e53688c23c5782022b8bcf6012a9a" Dec 09 16:50:04 crc kubenswrapper[4894]: I1209 16:50:04.853311 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-pn9dz/must-gather-xjnpp" Dec 09 16:50:04 crc kubenswrapper[4894]: I1209 16:50:04.874962 4894 scope.go:117] "RemoveContainer" containerID="4d585bb631c76873f173ca309f0f4ef760b09bc83ba0e5ad2fdda994558fd24b" Dec 09 16:50:06 crc kubenswrapper[4894]: I1209 16:50:06.121396 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="803d496d-7f84-4100-9f68-e2413ff22a54" path="/var/lib/kubelet/pods/803d496d-7f84-4100-9f68-e2413ff22a54/volumes" Dec 09 16:50:11 crc kubenswrapper[4894]: I1209 16:50:11.137425 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-fgxpp"] Dec 09 16:50:11 crc kubenswrapper[4894]: E1209 16:50:11.138581 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="803d496d-7f84-4100-9f68-e2413ff22a54" containerName="gather" Dec 09 16:50:11 crc kubenswrapper[4894]: I1209 16:50:11.138598 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="803d496d-7f84-4100-9f68-e2413ff22a54" containerName="gather" Dec 09 16:50:11 crc kubenswrapper[4894]: E1209 16:50:11.138661 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="803d496d-7f84-4100-9f68-e2413ff22a54" containerName="copy" Dec 09 16:50:11 crc kubenswrapper[4894]: I1209 16:50:11.138671 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="803d496d-7f84-4100-9f68-e2413ff22a54" containerName="copy" Dec 09 16:50:11 crc kubenswrapper[4894]: E1209 16:50:11.138692 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64eb0ebe-7743-4d2d-ba38-84ebb1b004e4" containerName="container-00" Dec 09 16:50:11 crc kubenswrapper[4894]: I1209 16:50:11.138702 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="64eb0ebe-7743-4d2d-ba38-84ebb1b004e4" containerName="container-00" Dec 09 16:50:11 crc kubenswrapper[4894]: I1209 16:50:11.139075 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="803d496d-7f84-4100-9f68-e2413ff22a54" containerName="gather" Dec 09 16:50:11 crc kubenswrapper[4894]: I1209 16:50:11.139104 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="803d496d-7f84-4100-9f68-e2413ff22a54" containerName="copy" Dec 09 16:50:11 crc kubenswrapper[4894]: I1209 16:50:11.139117 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="64eb0ebe-7743-4d2d-ba38-84ebb1b004e4" containerName="container-00" Dec 09 16:50:11 crc kubenswrapper[4894]: I1209 16:50:11.141352 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fgxpp" Dec 09 16:50:11 crc kubenswrapper[4894]: I1209 16:50:11.158983 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fgxpp"] Dec 09 16:50:11 crc kubenswrapper[4894]: I1209 16:50:11.282777 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bpmlr\" (UniqueName: \"kubernetes.io/projected/b7c79c9c-7993-4b48-bdea-357c1645edf4-kube-api-access-bpmlr\") pod \"redhat-operators-fgxpp\" (UID: \"b7c79c9c-7993-4b48-bdea-357c1645edf4\") " pod="openshift-marketplace/redhat-operators-fgxpp" Dec 09 16:50:11 crc kubenswrapper[4894]: I1209 16:50:11.282833 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7c79c9c-7993-4b48-bdea-357c1645edf4-utilities\") pod \"redhat-operators-fgxpp\" (UID: \"b7c79c9c-7993-4b48-bdea-357c1645edf4\") " pod="openshift-marketplace/redhat-operators-fgxpp" Dec 09 16:50:11 crc kubenswrapper[4894]: I1209 16:50:11.282856 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7c79c9c-7993-4b48-bdea-357c1645edf4-catalog-content\") pod \"redhat-operators-fgxpp\" (UID: \"b7c79c9c-7993-4b48-bdea-357c1645edf4\") " pod="openshift-marketplace/redhat-operators-fgxpp" Dec 09 16:50:11 crc kubenswrapper[4894]: I1209 16:50:11.385831 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bpmlr\" (UniqueName: \"kubernetes.io/projected/b7c79c9c-7993-4b48-bdea-357c1645edf4-kube-api-access-bpmlr\") pod \"redhat-operators-fgxpp\" (UID: \"b7c79c9c-7993-4b48-bdea-357c1645edf4\") " pod="openshift-marketplace/redhat-operators-fgxpp" Dec 09 16:50:11 crc kubenswrapper[4894]: I1209 16:50:11.385890 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7c79c9c-7993-4b48-bdea-357c1645edf4-utilities\") pod \"redhat-operators-fgxpp\" (UID: \"b7c79c9c-7993-4b48-bdea-357c1645edf4\") " pod="openshift-marketplace/redhat-operators-fgxpp" Dec 09 16:50:11 crc kubenswrapper[4894]: I1209 16:50:11.385915 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7c79c9c-7993-4b48-bdea-357c1645edf4-catalog-content\") pod \"redhat-operators-fgxpp\" (UID: \"b7c79c9c-7993-4b48-bdea-357c1645edf4\") " pod="openshift-marketplace/redhat-operators-fgxpp" Dec 09 16:50:11 crc kubenswrapper[4894]: I1209 16:50:11.386476 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7c79c9c-7993-4b48-bdea-357c1645edf4-utilities\") pod \"redhat-operators-fgxpp\" (UID: \"b7c79c9c-7993-4b48-bdea-357c1645edf4\") " pod="openshift-marketplace/redhat-operators-fgxpp" Dec 09 16:50:11 crc kubenswrapper[4894]: I1209 16:50:11.386633 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7c79c9c-7993-4b48-bdea-357c1645edf4-catalog-content\") pod \"redhat-operators-fgxpp\" (UID: \"b7c79c9c-7993-4b48-bdea-357c1645edf4\") " pod="openshift-marketplace/redhat-operators-fgxpp" Dec 09 16:50:11 crc kubenswrapper[4894]: I1209 16:50:11.407301 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bpmlr\" (UniqueName: \"kubernetes.io/projected/b7c79c9c-7993-4b48-bdea-357c1645edf4-kube-api-access-bpmlr\") pod \"redhat-operators-fgxpp\" (UID: \"b7c79c9c-7993-4b48-bdea-357c1645edf4\") " pod="openshift-marketplace/redhat-operators-fgxpp" Dec 09 16:50:11 crc kubenswrapper[4894]: I1209 16:50:11.462757 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fgxpp" Dec 09 16:50:11 crc kubenswrapper[4894]: I1209 16:50:11.970451 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-fgxpp"] Dec 09 16:50:12 crc kubenswrapper[4894]: I1209 16:50:12.930365 4894 generic.go:334] "Generic (PLEG): container finished" podID="b7c79c9c-7993-4b48-bdea-357c1645edf4" containerID="122062fcec484a9adadd4690918ad435a0bf4b0dc9919496be62c3d77bcf3e11" exitCode=0 Dec 09 16:50:12 crc kubenswrapper[4894]: I1209 16:50:12.930429 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fgxpp" event={"ID":"b7c79c9c-7993-4b48-bdea-357c1645edf4","Type":"ContainerDied","Data":"122062fcec484a9adadd4690918ad435a0bf4b0dc9919496be62c3d77bcf3e11"} Dec 09 16:50:12 crc kubenswrapper[4894]: I1209 16:50:12.930965 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fgxpp" event={"ID":"b7c79c9c-7993-4b48-bdea-357c1645edf4","Type":"ContainerStarted","Data":"f6b31fb117c9f37a4ed75df5904ea411063ee29ce2163fb75d83d3f668b7d9dc"} Dec 09 16:50:12 crc kubenswrapper[4894]: I1209 16:50:12.933799 4894 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Dec 09 16:50:14 crc kubenswrapper[4894]: I1209 16:50:14.958866 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fgxpp" event={"ID":"b7c79c9c-7993-4b48-bdea-357c1645edf4","Type":"ContainerStarted","Data":"eeb45e31f186d0deb1a42fd12198eaaa99255d0ec64393079860239852a1d9bc"} Dec 09 16:50:16 crc kubenswrapper[4894]: I1209 16:50:16.980382 4894 generic.go:334] "Generic (PLEG): container finished" podID="b7c79c9c-7993-4b48-bdea-357c1645edf4" containerID="eeb45e31f186d0deb1a42fd12198eaaa99255d0ec64393079860239852a1d9bc" exitCode=0 Dec 09 16:50:16 crc kubenswrapper[4894]: I1209 16:50:16.980479 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fgxpp" event={"ID":"b7c79c9c-7993-4b48-bdea-357c1645edf4","Type":"ContainerDied","Data":"eeb45e31f186d0deb1a42fd12198eaaa99255d0ec64393079860239852a1d9bc"} Dec 09 16:50:17 crc kubenswrapper[4894]: I1209 16:50:17.991353 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fgxpp" event={"ID":"b7c79c9c-7993-4b48-bdea-357c1645edf4","Type":"ContainerStarted","Data":"b99a04899b041bac0ff45fae79e4429712e9d25be8819c92c32d045309f69883"} Dec 09 16:50:18 crc kubenswrapper[4894]: I1209 16:50:18.017470 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-fgxpp" podStartSLOduration=2.330651536 podStartE2EDuration="7.017452181s" podCreationTimestamp="2025-12-09 16:50:11 +0000 UTC" firstStartedPulling="2025-12-09 16:50:12.933534439 +0000 UTC m=+4707.252745108" lastFinishedPulling="2025-12-09 16:50:17.620335074 +0000 UTC m=+4711.939545753" observedRunningTime="2025-12-09 16:50:18.008069127 +0000 UTC m=+4712.327279806" watchObservedRunningTime="2025-12-09 16:50:18.017452181 +0000 UTC m=+4712.336662850" Dec 09 16:50:21 crc kubenswrapper[4894]: I1209 16:50:21.462986 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-fgxpp" Dec 09 16:50:21 crc kubenswrapper[4894]: I1209 16:50:21.464615 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-fgxpp" Dec 09 16:50:22 crc kubenswrapper[4894]: I1209 16:50:22.514998 4894 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-fgxpp" podUID="b7c79c9c-7993-4b48-bdea-357c1645edf4" containerName="registry-server" probeResult="failure" output=< Dec 09 16:50:22 crc kubenswrapper[4894]: timeout: failed to connect service ":50051" within 1s Dec 09 16:50:22 crc kubenswrapper[4894]: > Dec 09 16:50:31 crc kubenswrapper[4894]: I1209 16:50:31.534822 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-fgxpp" Dec 09 16:50:31 crc kubenswrapper[4894]: I1209 16:50:31.626311 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-fgxpp" Dec 09 16:50:31 crc kubenswrapper[4894]: I1209 16:50:31.780245 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fgxpp"] Dec 09 16:50:33 crc kubenswrapper[4894]: I1209 16:50:33.132452 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-fgxpp" podUID="b7c79c9c-7993-4b48-bdea-357c1645edf4" containerName="registry-server" containerID="cri-o://b99a04899b041bac0ff45fae79e4429712e9d25be8819c92c32d045309f69883" gracePeriod=2 Dec 09 16:50:33 crc kubenswrapper[4894]: I1209 16:50:33.837357 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fgxpp" Dec 09 16:50:33 crc kubenswrapper[4894]: I1209 16:50:33.997722 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7c79c9c-7993-4b48-bdea-357c1645edf4-catalog-content\") pod \"b7c79c9c-7993-4b48-bdea-357c1645edf4\" (UID: \"b7c79c9c-7993-4b48-bdea-357c1645edf4\") " Dec 09 16:50:33 crc kubenswrapper[4894]: I1209 16:50:33.998137 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bpmlr\" (UniqueName: \"kubernetes.io/projected/b7c79c9c-7993-4b48-bdea-357c1645edf4-kube-api-access-bpmlr\") pod \"b7c79c9c-7993-4b48-bdea-357c1645edf4\" (UID: \"b7c79c9c-7993-4b48-bdea-357c1645edf4\") " Dec 09 16:50:33 crc kubenswrapper[4894]: I1209 16:50:33.998244 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7c79c9c-7993-4b48-bdea-357c1645edf4-utilities\") pod \"b7c79c9c-7993-4b48-bdea-357c1645edf4\" (UID: \"b7c79c9c-7993-4b48-bdea-357c1645edf4\") " Dec 09 16:50:33 crc kubenswrapper[4894]: I1209 16:50:33.999032 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7c79c9c-7993-4b48-bdea-357c1645edf4-utilities" (OuterVolumeSpecName: "utilities") pod "b7c79c9c-7993-4b48-bdea-357c1645edf4" (UID: "b7c79c9c-7993-4b48-bdea-357c1645edf4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:50:34 crc kubenswrapper[4894]: I1209 16:50:34.006176 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b7c79c9c-7993-4b48-bdea-357c1645edf4-kube-api-access-bpmlr" (OuterVolumeSpecName: "kube-api-access-bpmlr") pod "b7c79c9c-7993-4b48-bdea-357c1645edf4" (UID: "b7c79c9c-7993-4b48-bdea-357c1645edf4"). InnerVolumeSpecName "kube-api-access-bpmlr". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:50:34 crc kubenswrapper[4894]: I1209 16:50:34.101275 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bpmlr\" (UniqueName: \"kubernetes.io/projected/b7c79c9c-7993-4b48-bdea-357c1645edf4-kube-api-access-bpmlr\") on node \"crc\" DevicePath \"\"" Dec 09 16:50:34 crc kubenswrapper[4894]: I1209 16:50:34.101700 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b7c79c9c-7993-4b48-bdea-357c1645edf4-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:50:34 crc kubenswrapper[4894]: I1209 16:50:34.105018 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b7c79c9c-7993-4b48-bdea-357c1645edf4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b7c79c9c-7993-4b48-bdea-357c1645edf4" (UID: "b7c79c9c-7993-4b48-bdea-357c1645edf4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:50:34 crc kubenswrapper[4894]: I1209 16:50:34.141839 4894 generic.go:334] "Generic (PLEG): container finished" podID="b7c79c9c-7993-4b48-bdea-357c1645edf4" containerID="b99a04899b041bac0ff45fae79e4429712e9d25be8819c92c32d045309f69883" exitCode=0 Dec 09 16:50:34 crc kubenswrapper[4894]: I1209 16:50:34.141880 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fgxpp" event={"ID":"b7c79c9c-7993-4b48-bdea-357c1645edf4","Type":"ContainerDied","Data":"b99a04899b041bac0ff45fae79e4429712e9d25be8819c92c32d045309f69883"} Dec 09 16:50:34 crc kubenswrapper[4894]: I1209 16:50:34.141905 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-fgxpp" event={"ID":"b7c79c9c-7993-4b48-bdea-357c1645edf4","Type":"ContainerDied","Data":"f6b31fb117c9f37a4ed75df5904ea411063ee29ce2163fb75d83d3f668b7d9dc"} Dec 09 16:50:34 crc kubenswrapper[4894]: I1209 16:50:34.141922 4894 scope.go:117] "RemoveContainer" containerID="b99a04899b041bac0ff45fae79e4429712e9d25be8819c92c32d045309f69883" Dec 09 16:50:34 crc kubenswrapper[4894]: I1209 16:50:34.142029 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-fgxpp" Dec 09 16:50:34 crc kubenswrapper[4894]: I1209 16:50:34.169441 4894 scope.go:117] "RemoveContainer" containerID="eeb45e31f186d0deb1a42fd12198eaaa99255d0ec64393079860239852a1d9bc" Dec 09 16:50:34 crc kubenswrapper[4894]: I1209 16:50:34.179540 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-fgxpp"] Dec 09 16:50:34 crc kubenswrapper[4894]: I1209 16:50:34.188282 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-fgxpp"] Dec 09 16:50:34 crc kubenswrapper[4894]: I1209 16:50:34.203487 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b7c79c9c-7993-4b48-bdea-357c1645edf4-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:50:34 crc kubenswrapper[4894]: I1209 16:50:34.825698 4894 scope.go:117] "RemoveContainer" containerID="122062fcec484a9adadd4690918ad435a0bf4b0dc9919496be62c3d77bcf3e11" Dec 09 16:50:34 crc kubenswrapper[4894]: I1209 16:50:34.882081 4894 scope.go:117] "RemoveContainer" containerID="b99a04899b041bac0ff45fae79e4429712e9d25be8819c92c32d045309f69883" Dec 09 16:50:34 crc kubenswrapper[4894]: E1209 16:50:34.882470 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b99a04899b041bac0ff45fae79e4429712e9d25be8819c92c32d045309f69883\": container with ID starting with b99a04899b041bac0ff45fae79e4429712e9d25be8819c92c32d045309f69883 not found: ID does not exist" containerID="b99a04899b041bac0ff45fae79e4429712e9d25be8819c92c32d045309f69883" Dec 09 16:50:34 crc kubenswrapper[4894]: I1209 16:50:34.882541 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b99a04899b041bac0ff45fae79e4429712e9d25be8819c92c32d045309f69883"} err="failed to get container status \"b99a04899b041bac0ff45fae79e4429712e9d25be8819c92c32d045309f69883\": rpc error: code = NotFound desc = could not find container \"b99a04899b041bac0ff45fae79e4429712e9d25be8819c92c32d045309f69883\": container with ID starting with b99a04899b041bac0ff45fae79e4429712e9d25be8819c92c32d045309f69883 not found: ID does not exist" Dec 09 16:50:34 crc kubenswrapper[4894]: I1209 16:50:34.882572 4894 scope.go:117] "RemoveContainer" containerID="eeb45e31f186d0deb1a42fd12198eaaa99255d0ec64393079860239852a1d9bc" Dec 09 16:50:34 crc kubenswrapper[4894]: E1209 16:50:34.882889 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eeb45e31f186d0deb1a42fd12198eaaa99255d0ec64393079860239852a1d9bc\": container with ID starting with eeb45e31f186d0deb1a42fd12198eaaa99255d0ec64393079860239852a1d9bc not found: ID does not exist" containerID="eeb45e31f186d0deb1a42fd12198eaaa99255d0ec64393079860239852a1d9bc" Dec 09 16:50:34 crc kubenswrapper[4894]: I1209 16:50:34.882917 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eeb45e31f186d0deb1a42fd12198eaaa99255d0ec64393079860239852a1d9bc"} err="failed to get container status \"eeb45e31f186d0deb1a42fd12198eaaa99255d0ec64393079860239852a1d9bc\": rpc error: code = NotFound desc = could not find container \"eeb45e31f186d0deb1a42fd12198eaaa99255d0ec64393079860239852a1d9bc\": container with ID starting with eeb45e31f186d0deb1a42fd12198eaaa99255d0ec64393079860239852a1d9bc not found: ID does not exist" Dec 09 16:50:34 crc kubenswrapper[4894]: I1209 16:50:34.882938 4894 scope.go:117] "RemoveContainer" containerID="122062fcec484a9adadd4690918ad435a0bf4b0dc9919496be62c3d77bcf3e11" Dec 09 16:50:34 crc kubenswrapper[4894]: E1209 16:50:34.883209 4894 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"122062fcec484a9adadd4690918ad435a0bf4b0dc9919496be62c3d77bcf3e11\": container with ID starting with 122062fcec484a9adadd4690918ad435a0bf4b0dc9919496be62c3d77bcf3e11 not found: ID does not exist" containerID="122062fcec484a9adadd4690918ad435a0bf4b0dc9919496be62c3d77bcf3e11" Dec 09 16:50:34 crc kubenswrapper[4894]: I1209 16:50:34.883249 4894 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"122062fcec484a9adadd4690918ad435a0bf4b0dc9919496be62c3d77bcf3e11"} err="failed to get container status \"122062fcec484a9adadd4690918ad435a0bf4b0dc9919496be62c3d77bcf3e11\": rpc error: code = NotFound desc = could not find container \"122062fcec484a9adadd4690918ad435a0bf4b0dc9919496be62c3d77bcf3e11\": container with ID starting with 122062fcec484a9adadd4690918ad435a0bf4b0dc9919496be62c3d77bcf3e11 not found: ID does not exist" Dec 09 16:50:36 crc kubenswrapper[4894]: I1209 16:50:36.118816 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b7c79c9c-7993-4b48-bdea-357c1645edf4" path="/var/lib/kubelet/pods/b7c79c9c-7993-4b48-bdea-357c1645edf4/volumes" Dec 09 16:50:53 crc kubenswrapper[4894]: E1209 16:50:53.107194 4894 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 09 16:50:58 crc kubenswrapper[4894]: I1209 16:50:58.966296 4894 scope.go:117] "RemoveContainer" containerID="4b086c2991f9d8a7a0c710e444cb2ac0ad7d6dd29a053eff10716040119f9cc0" Dec 09 16:51:26 crc kubenswrapper[4894]: I1209 16:51:26.690699 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-wpfnw"] Dec 09 16:51:26 crc kubenswrapper[4894]: E1209 16:51:26.691733 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7c79c9c-7993-4b48-bdea-357c1645edf4" containerName="registry-server" Dec 09 16:51:26 crc kubenswrapper[4894]: I1209 16:51:26.691749 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7c79c9c-7993-4b48-bdea-357c1645edf4" containerName="registry-server" Dec 09 16:51:26 crc kubenswrapper[4894]: E1209 16:51:26.691786 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7c79c9c-7993-4b48-bdea-357c1645edf4" containerName="extract-content" Dec 09 16:51:26 crc kubenswrapper[4894]: I1209 16:51:26.691795 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7c79c9c-7993-4b48-bdea-357c1645edf4" containerName="extract-content" Dec 09 16:51:26 crc kubenswrapper[4894]: E1209 16:51:26.691815 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b7c79c9c-7993-4b48-bdea-357c1645edf4" containerName="extract-utilities" Dec 09 16:51:26 crc kubenswrapper[4894]: I1209 16:51:26.691826 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="b7c79c9c-7993-4b48-bdea-357c1645edf4" containerName="extract-utilities" Dec 09 16:51:26 crc kubenswrapper[4894]: I1209 16:51:26.692121 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="b7c79c9c-7993-4b48-bdea-357c1645edf4" containerName="registry-server" Dec 09 16:51:26 crc kubenswrapper[4894]: I1209 16:51:26.693885 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wpfnw" Dec 09 16:51:26 crc kubenswrapper[4894]: I1209 16:51:26.710114 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wpfnw"] Dec 09 16:51:26 crc kubenswrapper[4894]: I1209 16:51:26.789571 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mhrst\" (UniqueName: \"kubernetes.io/projected/8fe84948-a66d-4207-9d6a-b50ad652fea8-kube-api-access-mhrst\") pod \"certified-operators-wpfnw\" (UID: \"8fe84948-a66d-4207-9d6a-b50ad652fea8\") " pod="openshift-marketplace/certified-operators-wpfnw" Dec 09 16:51:26 crc kubenswrapper[4894]: I1209 16:51:26.789899 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe84948-a66d-4207-9d6a-b50ad652fea8-utilities\") pod \"certified-operators-wpfnw\" (UID: \"8fe84948-a66d-4207-9d6a-b50ad652fea8\") " pod="openshift-marketplace/certified-operators-wpfnw" Dec 09 16:51:26 crc kubenswrapper[4894]: I1209 16:51:26.789965 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe84948-a66d-4207-9d6a-b50ad652fea8-catalog-content\") pod \"certified-operators-wpfnw\" (UID: \"8fe84948-a66d-4207-9d6a-b50ad652fea8\") " pod="openshift-marketplace/certified-operators-wpfnw" Dec 09 16:51:26 crc kubenswrapper[4894]: I1209 16:51:26.890704 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe84948-a66d-4207-9d6a-b50ad652fea8-utilities\") pod \"certified-operators-wpfnw\" (UID: \"8fe84948-a66d-4207-9d6a-b50ad652fea8\") " pod="openshift-marketplace/certified-operators-wpfnw" Dec 09 16:51:26 crc kubenswrapper[4894]: I1209 16:51:26.890785 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe84948-a66d-4207-9d6a-b50ad652fea8-catalog-content\") pod \"certified-operators-wpfnw\" (UID: \"8fe84948-a66d-4207-9d6a-b50ad652fea8\") " pod="openshift-marketplace/certified-operators-wpfnw" Dec 09 16:51:26 crc kubenswrapper[4894]: I1209 16:51:26.890820 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mhrst\" (UniqueName: \"kubernetes.io/projected/8fe84948-a66d-4207-9d6a-b50ad652fea8-kube-api-access-mhrst\") pod \"certified-operators-wpfnw\" (UID: \"8fe84948-a66d-4207-9d6a-b50ad652fea8\") " pod="openshift-marketplace/certified-operators-wpfnw" Dec 09 16:51:26 crc kubenswrapper[4894]: I1209 16:51:26.891525 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8fe84948-a66d-4207-9d6a-b50ad652fea8-utilities\") pod \"certified-operators-wpfnw\" (UID: \"8fe84948-a66d-4207-9d6a-b50ad652fea8\") " pod="openshift-marketplace/certified-operators-wpfnw" Dec 09 16:51:26 crc kubenswrapper[4894]: I1209 16:51:26.891606 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8fe84948-a66d-4207-9d6a-b50ad652fea8-catalog-content\") pod \"certified-operators-wpfnw\" (UID: \"8fe84948-a66d-4207-9d6a-b50ad652fea8\") " pod="openshift-marketplace/certified-operators-wpfnw" Dec 09 16:51:26 crc kubenswrapper[4894]: I1209 16:51:26.911091 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mhrst\" (UniqueName: \"kubernetes.io/projected/8fe84948-a66d-4207-9d6a-b50ad652fea8-kube-api-access-mhrst\") pod \"certified-operators-wpfnw\" (UID: \"8fe84948-a66d-4207-9d6a-b50ad652fea8\") " pod="openshift-marketplace/certified-operators-wpfnw" Dec 09 16:51:27 crc kubenswrapper[4894]: I1209 16:51:27.015896 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-wpfnw" Dec 09 16:51:27 crc kubenswrapper[4894]: I1209 16:51:27.531280 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wpfnw"] Dec 09 16:51:27 crc kubenswrapper[4894]: I1209 16:51:27.670741 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wpfnw" event={"ID":"8fe84948-a66d-4207-9d6a-b50ad652fea8","Type":"ContainerStarted","Data":"36621c03bfca2d7826cccfdb5a0c566fbdca0d550d1f550995822bbbe5ac4f94"} Dec 09 16:51:28 crc kubenswrapper[4894]: I1209 16:51:28.681144 4894 generic.go:334] "Generic (PLEG): container finished" podID="8fe84948-a66d-4207-9d6a-b50ad652fea8" containerID="fc4e4f753fc5a74be20b232d7a680f32e292549056984f336251ff9f6ff76fd0" exitCode=0 Dec 09 16:51:28 crc kubenswrapper[4894]: I1209 16:51:28.681186 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wpfnw" event={"ID":"8fe84948-a66d-4207-9d6a-b50ad652fea8","Type":"ContainerDied","Data":"fc4e4f753fc5a74be20b232d7a680f32e292549056984f336251ff9f6ff76fd0"} Dec 09 16:51:33 crc kubenswrapper[4894]: I1209 16:51:33.726821 4894 generic.go:334] "Generic (PLEG): container finished" podID="8fe84948-a66d-4207-9d6a-b50ad652fea8" containerID="fea917142b6d50aecb1d03db5efcf7ce9fe34b23d59285729e87d8b3071d2b3e" exitCode=0 Dec 09 16:51:33 crc kubenswrapper[4894]: I1209 16:51:33.726908 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wpfnw" event={"ID":"8fe84948-a66d-4207-9d6a-b50ad652fea8","Type":"ContainerDied","Data":"fea917142b6d50aecb1d03db5efcf7ce9fe34b23d59285729e87d8b3071d2b3e"} Dec 09 16:51:34 crc kubenswrapper[4894]: I1209 16:51:34.737860 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-wpfnw" event={"ID":"8fe84948-a66d-4207-9d6a-b50ad652fea8","Type":"ContainerStarted","Data":"84d365417970f97ee2cf5b14ddbb299ec940efb9bf518448a57303e97f7475d1"} Dec 09 16:51:34 crc kubenswrapper[4894]: I1209 16:51:34.759118 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-wpfnw" podStartSLOduration=3.257234042 podStartE2EDuration="8.759097438s" podCreationTimestamp="2025-12-09 16:51:26 +0000 UTC" firstStartedPulling="2025-12-09 16:51:28.684356721 +0000 UTC m=+4783.003567390" lastFinishedPulling="2025-12-09 16:51:34.186220077 +0000 UTC m=+4788.505430786" observedRunningTime="2025-12-09 16:51:34.753944219 +0000 UTC m=+4789.073154888" watchObservedRunningTime="2025-12-09 16:51:34.759097438 +0000 UTC m=+4789.078308107" Dec 09 16:51:37 crc kubenswrapper[4894]: I1209 16:51:37.016043 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-wpfnw" Dec 09 16:51:37 crc kubenswrapper[4894]: I1209 16:51:37.016444 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-wpfnw" Dec 09 16:51:37 crc kubenswrapper[4894]: I1209 16:51:37.063436 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-wpfnw" Dec 09 16:51:42 crc kubenswrapper[4894]: I1209 16:51:42.257871 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:51:42 crc kubenswrapper[4894]: I1209 16:51:42.258395 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:51:47 crc kubenswrapper[4894]: I1209 16:51:47.485558 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-wpfnw" Dec 09 16:51:47 crc kubenswrapper[4894]: I1209 16:51:47.547023 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-wpfnw"] Dec 09 16:51:47 crc kubenswrapper[4894]: I1209 16:51:47.592001 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lnlst"] Dec 09 16:51:47 crc kubenswrapper[4894]: I1209 16:51:47.592215 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-lnlst" podUID="74ee2091-d89d-4945-af03-5b9887bdfa24" containerName="registry-server" containerID="cri-o://9d0d808aaf5c6a7d109a94a5530688e8d06c0b619317ad4409e3d0248e473339" gracePeriod=2 Dec 09 16:51:47 crc kubenswrapper[4894]: I1209 16:51:47.855428 4894 generic.go:334] "Generic (PLEG): container finished" podID="74ee2091-d89d-4945-af03-5b9887bdfa24" containerID="9d0d808aaf5c6a7d109a94a5530688e8d06c0b619317ad4409e3d0248e473339" exitCode=0 Dec 09 16:51:47 crc kubenswrapper[4894]: I1209 16:51:47.855511 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnlst" event={"ID":"74ee2091-d89d-4945-af03-5b9887bdfa24","Type":"ContainerDied","Data":"9d0d808aaf5c6a7d109a94a5530688e8d06c0b619317ad4409e3d0248e473339"} Dec 09 16:51:48 crc kubenswrapper[4894]: I1209 16:51:48.077072 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lnlst" Dec 09 16:51:48 crc kubenswrapper[4894]: I1209 16:51:48.221714 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5sc2\" (UniqueName: \"kubernetes.io/projected/74ee2091-d89d-4945-af03-5b9887bdfa24-kube-api-access-x5sc2\") pod \"74ee2091-d89d-4945-af03-5b9887bdfa24\" (UID: \"74ee2091-d89d-4945-af03-5b9887bdfa24\") " Dec 09 16:51:48 crc kubenswrapper[4894]: I1209 16:51:48.221838 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74ee2091-d89d-4945-af03-5b9887bdfa24-catalog-content\") pod \"74ee2091-d89d-4945-af03-5b9887bdfa24\" (UID: \"74ee2091-d89d-4945-af03-5b9887bdfa24\") " Dec 09 16:51:48 crc kubenswrapper[4894]: I1209 16:51:48.221885 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74ee2091-d89d-4945-af03-5b9887bdfa24-utilities\") pod \"74ee2091-d89d-4945-af03-5b9887bdfa24\" (UID: \"74ee2091-d89d-4945-af03-5b9887bdfa24\") " Dec 09 16:51:48 crc kubenswrapper[4894]: I1209 16:51:48.222345 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74ee2091-d89d-4945-af03-5b9887bdfa24-utilities" (OuterVolumeSpecName: "utilities") pod "74ee2091-d89d-4945-af03-5b9887bdfa24" (UID: "74ee2091-d89d-4945-af03-5b9887bdfa24"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:51:48 crc kubenswrapper[4894]: I1209 16:51:48.229090 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/74ee2091-d89d-4945-af03-5b9887bdfa24-kube-api-access-x5sc2" (OuterVolumeSpecName: "kube-api-access-x5sc2") pod "74ee2091-d89d-4945-af03-5b9887bdfa24" (UID: "74ee2091-d89d-4945-af03-5b9887bdfa24"). InnerVolumeSpecName "kube-api-access-x5sc2". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:51:48 crc kubenswrapper[4894]: I1209 16:51:48.269920 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/74ee2091-d89d-4945-af03-5b9887bdfa24-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "74ee2091-d89d-4945-af03-5b9887bdfa24" (UID: "74ee2091-d89d-4945-af03-5b9887bdfa24"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:51:48 crc kubenswrapper[4894]: I1209 16:51:48.324029 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5sc2\" (UniqueName: \"kubernetes.io/projected/74ee2091-d89d-4945-af03-5b9887bdfa24-kube-api-access-x5sc2\") on node \"crc\" DevicePath \"\"" Dec 09 16:51:48 crc kubenswrapper[4894]: I1209 16:51:48.324062 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/74ee2091-d89d-4945-af03-5b9887bdfa24-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:51:48 crc kubenswrapper[4894]: I1209 16:51:48.324072 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/74ee2091-d89d-4945-af03-5b9887bdfa24-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:51:48 crc kubenswrapper[4894]: I1209 16:51:48.868923 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-lnlst" event={"ID":"74ee2091-d89d-4945-af03-5b9887bdfa24","Type":"ContainerDied","Data":"df416025652fa71db66c630370b5eec8678d35aac24d11431e254bf98a468c91"} Dec 09 16:51:48 crc kubenswrapper[4894]: I1209 16:51:48.868976 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-lnlst" Dec 09 16:51:48 crc kubenswrapper[4894]: I1209 16:51:48.868987 4894 scope.go:117] "RemoveContainer" containerID="9d0d808aaf5c6a7d109a94a5530688e8d06c0b619317ad4409e3d0248e473339" Dec 09 16:51:48 crc kubenswrapper[4894]: I1209 16:51:48.910825 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-lnlst"] Dec 09 16:51:48 crc kubenswrapper[4894]: I1209 16:51:48.916487 4894 scope.go:117] "RemoveContainer" containerID="5b72cfefd9241bdc06445f6dcf80c5109c59a376b3e91843424f3c83ef361823" Dec 09 16:51:48 crc kubenswrapper[4894]: I1209 16:51:48.919364 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-lnlst"] Dec 09 16:51:48 crc kubenswrapper[4894]: I1209 16:51:48.956962 4894 scope.go:117] "RemoveContainer" containerID="f63d61cd8a843613c1d00a8dc44375294a547d79110e11143738101cdd201a65" Dec 09 16:51:50 crc kubenswrapper[4894]: I1209 16:51:50.121430 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="74ee2091-d89d-4945-af03-5b9887bdfa24" path="/var/lib/kubelet/pods/74ee2091-d89d-4945-af03-5b9887bdfa24/volumes" Dec 09 16:52:01 crc kubenswrapper[4894]: I1209 16:52:01.521563 4894 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6r9rq"] Dec 09 16:52:01 crc kubenswrapper[4894]: E1209 16:52:01.522581 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74ee2091-d89d-4945-af03-5b9887bdfa24" containerName="registry-server" Dec 09 16:52:01 crc kubenswrapper[4894]: I1209 16:52:01.522599 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="74ee2091-d89d-4945-af03-5b9887bdfa24" containerName="registry-server" Dec 09 16:52:01 crc kubenswrapper[4894]: E1209 16:52:01.522629 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74ee2091-d89d-4945-af03-5b9887bdfa24" containerName="extract-utilities" Dec 09 16:52:01 crc kubenswrapper[4894]: I1209 16:52:01.522664 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="74ee2091-d89d-4945-af03-5b9887bdfa24" containerName="extract-utilities" Dec 09 16:52:01 crc kubenswrapper[4894]: E1209 16:52:01.522742 4894 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="74ee2091-d89d-4945-af03-5b9887bdfa24" containerName="extract-content" Dec 09 16:52:01 crc kubenswrapper[4894]: I1209 16:52:01.522749 4894 state_mem.go:107] "Deleted CPUSet assignment" podUID="74ee2091-d89d-4945-af03-5b9887bdfa24" containerName="extract-content" Dec 09 16:52:01 crc kubenswrapper[4894]: I1209 16:52:01.522946 4894 memory_manager.go:354] "RemoveStaleState removing state" podUID="74ee2091-d89d-4945-af03-5b9887bdfa24" containerName="registry-server" Dec 09 16:52:01 crc kubenswrapper[4894]: I1209 16:52:01.527799 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6r9rq" Dec 09 16:52:01 crc kubenswrapper[4894]: I1209 16:52:01.537768 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6r9rq"] Dec 09 16:52:01 crc kubenswrapper[4894]: I1209 16:52:01.583897 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ph5q7\" (UniqueName: \"kubernetes.io/projected/19a0a66c-b085-4fe3-861a-fa479782b58e-kube-api-access-ph5q7\") pod \"community-operators-6r9rq\" (UID: \"19a0a66c-b085-4fe3-861a-fa479782b58e\") " pod="openshift-marketplace/community-operators-6r9rq" Dec 09 16:52:01 crc kubenswrapper[4894]: I1209 16:52:01.583950 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19a0a66c-b085-4fe3-861a-fa479782b58e-utilities\") pod \"community-operators-6r9rq\" (UID: \"19a0a66c-b085-4fe3-861a-fa479782b58e\") " pod="openshift-marketplace/community-operators-6r9rq" Dec 09 16:52:01 crc kubenswrapper[4894]: I1209 16:52:01.584036 4894 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19a0a66c-b085-4fe3-861a-fa479782b58e-catalog-content\") pod \"community-operators-6r9rq\" (UID: \"19a0a66c-b085-4fe3-861a-fa479782b58e\") " pod="openshift-marketplace/community-operators-6r9rq" Dec 09 16:52:01 crc kubenswrapper[4894]: I1209 16:52:01.685791 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ph5q7\" (UniqueName: \"kubernetes.io/projected/19a0a66c-b085-4fe3-861a-fa479782b58e-kube-api-access-ph5q7\") pod \"community-operators-6r9rq\" (UID: \"19a0a66c-b085-4fe3-861a-fa479782b58e\") " pod="openshift-marketplace/community-operators-6r9rq" Dec 09 16:52:01 crc kubenswrapper[4894]: I1209 16:52:01.685850 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19a0a66c-b085-4fe3-861a-fa479782b58e-utilities\") pod \"community-operators-6r9rq\" (UID: \"19a0a66c-b085-4fe3-861a-fa479782b58e\") " pod="openshift-marketplace/community-operators-6r9rq" Dec 09 16:52:01 crc kubenswrapper[4894]: I1209 16:52:01.685938 4894 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19a0a66c-b085-4fe3-861a-fa479782b58e-catalog-content\") pod \"community-operators-6r9rq\" (UID: \"19a0a66c-b085-4fe3-861a-fa479782b58e\") " pod="openshift-marketplace/community-operators-6r9rq" Dec 09 16:52:01 crc kubenswrapper[4894]: I1209 16:52:01.686415 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19a0a66c-b085-4fe3-861a-fa479782b58e-utilities\") pod \"community-operators-6r9rq\" (UID: \"19a0a66c-b085-4fe3-861a-fa479782b58e\") " pod="openshift-marketplace/community-operators-6r9rq" Dec 09 16:52:01 crc kubenswrapper[4894]: I1209 16:52:01.686479 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19a0a66c-b085-4fe3-861a-fa479782b58e-catalog-content\") pod \"community-operators-6r9rq\" (UID: \"19a0a66c-b085-4fe3-861a-fa479782b58e\") " pod="openshift-marketplace/community-operators-6r9rq" Dec 09 16:52:01 crc kubenswrapper[4894]: I1209 16:52:01.705698 4894 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ph5q7\" (UniqueName: \"kubernetes.io/projected/19a0a66c-b085-4fe3-861a-fa479782b58e-kube-api-access-ph5q7\") pod \"community-operators-6r9rq\" (UID: \"19a0a66c-b085-4fe3-861a-fa479782b58e\") " pod="openshift-marketplace/community-operators-6r9rq" Dec 09 16:52:01 crc kubenswrapper[4894]: I1209 16:52:01.848675 4894 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6r9rq" Dec 09 16:52:02 crc kubenswrapper[4894]: I1209 16:52:02.191730 4894 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6r9rq"] Dec 09 16:52:03 crc kubenswrapper[4894]: I1209 16:52:03.003656 4894 generic.go:334] "Generic (PLEG): container finished" podID="19a0a66c-b085-4fe3-861a-fa479782b58e" containerID="2b350665bc5a618a06e9ff245c57e8c9df67add38633e65d587f049bf61a2168" exitCode=0 Dec 09 16:52:03 crc kubenswrapper[4894]: I1209 16:52:03.003767 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6r9rq" event={"ID":"19a0a66c-b085-4fe3-861a-fa479782b58e","Type":"ContainerDied","Data":"2b350665bc5a618a06e9ff245c57e8c9df67add38633e65d587f049bf61a2168"} Dec 09 16:52:03 crc kubenswrapper[4894]: I1209 16:52:03.004532 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6r9rq" event={"ID":"19a0a66c-b085-4fe3-861a-fa479782b58e","Type":"ContainerStarted","Data":"c352565de9e6dbcb4349cd98423658fd5c301d1eb9d415cea49bc3dc675b926e"} Dec 09 16:52:06 crc kubenswrapper[4894]: I1209 16:52:06.029130 4894 generic.go:334] "Generic (PLEG): container finished" podID="19a0a66c-b085-4fe3-861a-fa479782b58e" containerID="12a85bfab978817fb36d039261a417f82167ebaa185e2c834c8d2b11f7817524" exitCode=0 Dec 09 16:52:06 crc kubenswrapper[4894]: I1209 16:52:06.030414 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6r9rq" event={"ID":"19a0a66c-b085-4fe3-861a-fa479782b58e","Type":"ContainerDied","Data":"12a85bfab978817fb36d039261a417f82167ebaa185e2c834c8d2b11f7817524"} Dec 09 16:52:07 crc kubenswrapper[4894]: I1209 16:52:07.043740 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6r9rq" event={"ID":"19a0a66c-b085-4fe3-861a-fa479782b58e","Type":"ContainerStarted","Data":"18124a5641bc72b1e13bdf018e22c72f2ec0e1706e9995d9bb65defeb2eeca10"} Dec 09 16:52:07 crc kubenswrapper[4894]: I1209 16:52:07.064923 4894 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6r9rq" podStartSLOduration=2.630905025 podStartE2EDuration="6.064904754s" podCreationTimestamp="2025-12-09 16:52:01 +0000 UTC" firstStartedPulling="2025-12-09 16:52:03.00607318 +0000 UTC m=+4817.325283849" lastFinishedPulling="2025-12-09 16:52:06.440072899 +0000 UTC m=+4820.759283578" observedRunningTime="2025-12-09 16:52:07.063542497 +0000 UTC m=+4821.382753166" watchObservedRunningTime="2025-12-09 16:52:07.064904754 +0000 UTC m=+4821.384115413" Dec 09 16:52:11 crc kubenswrapper[4894]: I1209 16:52:11.849763 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6r9rq" Dec 09 16:52:11 crc kubenswrapper[4894]: I1209 16:52:11.850255 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6r9rq" Dec 09 16:52:11 crc kubenswrapper[4894]: I1209 16:52:11.906817 4894 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6r9rq" Dec 09 16:52:12 crc kubenswrapper[4894]: I1209 16:52:12.153712 4894 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6r9rq" Dec 09 16:52:12 crc kubenswrapper[4894]: I1209 16:52:12.205879 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6r9rq"] Dec 09 16:52:12 crc kubenswrapper[4894]: I1209 16:52:12.257963 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:52:12 crc kubenswrapper[4894]: I1209 16:52:12.258037 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:52:14 crc kubenswrapper[4894]: I1209 16:52:14.105485 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6r9rq" podUID="19a0a66c-b085-4fe3-861a-fa479782b58e" containerName="registry-server" containerID="cri-o://18124a5641bc72b1e13bdf018e22c72f2ec0e1706e9995d9bb65defeb2eeca10" gracePeriod=2 Dec 09 16:52:15 crc kubenswrapper[4894]: I1209 16:52:15.117709 4894 generic.go:334] "Generic (PLEG): container finished" podID="19a0a66c-b085-4fe3-861a-fa479782b58e" containerID="18124a5641bc72b1e13bdf018e22c72f2ec0e1706e9995d9bb65defeb2eeca10" exitCode=0 Dec 09 16:52:15 crc kubenswrapper[4894]: I1209 16:52:15.117754 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6r9rq" event={"ID":"19a0a66c-b085-4fe3-861a-fa479782b58e","Type":"ContainerDied","Data":"18124a5641bc72b1e13bdf018e22c72f2ec0e1706e9995d9bb65defeb2eeca10"} Dec 09 16:52:15 crc kubenswrapper[4894]: I1209 16:52:15.814684 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6r9rq" Dec 09 16:52:15 crc kubenswrapper[4894]: I1209 16:52:15.880818 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ph5q7\" (UniqueName: \"kubernetes.io/projected/19a0a66c-b085-4fe3-861a-fa479782b58e-kube-api-access-ph5q7\") pod \"19a0a66c-b085-4fe3-861a-fa479782b58e\" (UID: \"19a0a66c-b085-4fe3-861a-fa479782b58e\") " Dec 09 16:52:15 crc kubenswrapper[4894]: I1209 16:52:15.881092 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19a0a66c-b085-4fe3-861a-fa479782b58e-utilities\") pod \"19a0a66c-b085-4fe3-861a-fa479782b58e\" (UID: \"19a0a66c-b085-4fe3-861a-fa479782b58e\") " Dec 09 16:52:15 crc kubenswrapper[4894]: I1209 16:52:15.881152 4894 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19a0a66c-b085-4fe3-861a-fa479782b58e-catalog-content\") pod \"19a0a66c-b085-4fe3-861a-fa479782b58e\" (UID: \"19a0a66c-b085-4fe3-861a-fa479782b58e\") " Dec 09 16:52:15 crc kubenswrapper[4894]: I1209 16:52:15.882176 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19a0a66c-b085-4fe3-861a-fa479782b58e-utilities" (OuterVolumeSpecName: "utilities") pod "19a0a66c-b085-4fe3-861a-fa479782b58e" (UID: "19a0a66c-b085-4fe3-861a-fa479782b58e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:52:15 crc kubenswrapper[4894]: I1209 16:52:15.886721 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/19a0a66c-b085-4fe3-861a-fa479782b58e-kube-api-access-ph5q7" (OuterVolumeSpecName: "kube-api-access-ph5q7") pod "19a0a66c-b085-4fe3-861a-fa479782b58e" (UID: "19a0a66c-b085-4fe3-861a-fa479782b58e"). InnerVolumeSpecName "kube-api-access-ph5q7". PluginName "kubernetes.io/projected", VolumeGidValue "" Dec 09 16:52:15 crc kubenswrapper[4894]: I1209 16:52:15.932418 4894 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/19a0a66c-b085-4fe3-861a-fa479782b58e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "19a0a66c-b085-4fe3-861a-fa479782b58e" (UID: "19a0a66c-b085-4fe3-861a-fa479782b58e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Dec 09 16:52:15 crc kubenswrapper[4894]: I1209 16:52:15.983789 4894 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/19a0a66c-b085-4fe3-861a-fa479782b58e-utilities\") on node \"crc\" DevicePath \"\"" Dec 09 16:52:15 crc kubenswrapper[4894]: I1209 16:52:15.983829 4894 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/19a0a66c-b085-4fe3-861a-fa479782b58e-catalog-content\") on node \"crc\" DevicePath \"\"" Dec 09 16:52:15 crc kubenswrapper[4894]: I1209 16:52:15.983844 4894 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ph5q7\" (UniqueName: \"kubernetes.io/projected/19a0a66c-b085-4fe3-861a-fa479782b58e-kube-api-access-ph5q7\") on node \"crc\" DevicePath \"\"" Dec 09 16:52:16 crc kubenswrapper[4894]: I1209 16:52:16.130093 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6r9rq" event={"ID":"19a0a66c-b085-4fe3-861a-fa479782b58e","Type":"ContainerDied","Data":"c352565de9e6dbcb4349cd98423658fd5c301d1eb9d415cea49bc3dc675b926e"} Dec 09 16:52:16 crc kubenswrapper[4894]: I1209 16:52:16.130147 4894 scope.go:117] "RemoveContainer" containerID="18124a5641bc72b1e13bdf018e22c72f2ec0e1706e9995d9bb65defeb2eeca10" Dec 09 16:52:16 crc kubenswrapper[4894]: I1209 16:52:16.130178 4894 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6r9rq" Dec 09 16:52:16 crc kubenswrapper[4894]: I1209 16:52:16.150422 4894 scope.go:117] "RemoveContainer" containerID="12a85bfab978817fb36d039261a417f82167ebaa185e2c834c8d2b11f7817524" Dec 09 16:52:16 crc kubenswrapper[4894]: I1209 16:52:16.173568 4894 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6r9rq"] Dec 09 16:52:16 crc kubenswrapper[4894]: I1209 16:52:16.179088 4894 scope.go:117] "RemoveContainer" containerID="2b350665bc5a618a06e9ff245c57e8c9df67add38633e65d587f049bf61a2168" Dec 09 16:52:16 crc kubenswrapper[4894]: I1209 16:52:16.183710 4894 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6r9rq"] Dec 09 16:52:18 crc kubenswrapper[4894]: I1209 16:52:18.132238 4894 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="19a0a66c-b085-4fe3-861a-fa479782b58e" path="/var/lib/kubelet/pods/19a0a66c-b085-4fe3-861a-fa479782b58e/volumes" Dec 09 16:52:19 crc kubenswrapper[4894]: E1209 16:52:19.106343 4894 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 09 16:52:42 crc kubenswrapper[4894]: I1209 16:52:42.257508 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:52:42 crc kubenswrapper[4894]: I1209 16:52:42.258087 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:52:42 crc kubenswrapper[4894]: I1209 16:52:42.258129 4894 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" Dec 09 16:52:42 crc kubenswrapper[4894]: I1209 16:52:42.258604 4894 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d6834fb5fc2cfe31b11cb42928fcf576c01fe5393bc29ae78a328cdf98460675"} pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Dec 09 16:52:42 crc kubenswrapper[4894]: I1209 16:52:42.258688 4894 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" containerID="cri-o://d6834fb5fc2cfe31b11cb42928fcf576c01fe5393bc29ae78a328cdf98460675" gracePeriod=600 Dec 09 16:52:42 crc kubenswrapper[4894]: I1209 16:52:42.387141 4894 generic.go:334] "Generic (PLEG): container finished" podID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerID="d6834fb5fc2cfe31b11cb42928fcf576c01fe5393bc29ae78a328cdf98460675" exitCode=0 Dec 09 16:52:42 crc kubenswrapper[4894]: I1209 16:52:42.387206 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerDied","Data":"d6834fb5fc2cfe31b11cb42928fcf576c01fe5393bc29ae78a328cdf98460675"} Dec 09 16:52:42 crc kubenswrapper[4894]: I1209 16:52:42.387487 4894 scope.go:117] "RemoveContainer" containerID="30038f074f860cd1fb801200efaea5179ed0f2ecf9bda3dab96a2b9046ef0c01" Dec 09 16:52:43 crc kubenswrapper[4894]: I1209 16:52:43.398654 4894 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" event={"ID":"cfa97fcb-7e7c-4202-a2d0-d283fa49b112","Type":"ContainerStarted","Data":"627061283e47266f098a85597249f714a7594bdbc1dd4d718a45073db1e2e366"} Dec 09 16:53:39 crc kubenswrapper[4894]: E1209 16:53:39.106527 4894 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" Dec 09 16:54:42 crc kubenswrapper[4894]: I1209 16:54:42.257442 4894 patch_prober.go:28] interesting pod/machine-config-daemon-rj7cb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Dec 09 16:54:42 crc kubenswrapper[4894]: I1209 16:54:42.258110 4894 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-rj7cb" podUID="cfa97fcb-7e7c-4202-a2d0-d283fa49b112" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Dec 09 16:54:44 crc kubenswrapper[4894]: E1209 16:54:44.107878 4894 kubelet_pods.go:538] "Hostname for pod was too long, truncated it" podName="test-operator-logs-pod-horizontest-horizontest-tests-horizontest" hostnameMaxLen=63 truncatedHostname="test-operator-logs-pod-horizontest-horizontest-tests-horizontes" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515116052137024446 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015116052140017355 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015116040023016475 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015116040024015446 5ustar corecore